The Knowledge Project with Shane Parrish - Bret Taylor: A Vision for AI’s Next Frontier
Episode Date: April 15, 2025What happens when one of the most legendary minds in tech delves deep into the real workings of modern AI? A 2-hour long masterclass that you don’t want to miss. Bret Taylor, current chairman of ...OpenAI, unpacks why AI is transforming software engineering forever, how founders can survive acquisition (he’s done it twice), and why the true bottlenecks in AI aren’t what most think. Drawing on his extensive experiences at Facebook, Google, Twitter and more, he explains why the next phase of AI won’t just be about building better models—it’s about creating entirely new ways for us to work with them. Bret exposes the reality gap between what AI insiders understand and what everyone else believes. Listen now to recalibrate your thinking before your competitors do. (00:02:46) Aha Moments with AI (00:04:43) Founders Working for Founders (00:07:59) Acquisition Process (00:14:14) The Role of a Board (00:17:05) Founder Mode (00:20:29) Engineers as Leaders (00:24:54) Applying First Principles in Business (00:28:43) The Future of Software Engineering (00:35:11) Efficiency and Verification of AI-Generated Code (00:36:46) The Future of Software Development (00:37:24) Defining AGI (00:47:03) AI Self-Improvement? (00:47:58) Safety Measures and Supervision in AI (00:49:47) Benefiting Humanity and AI Safety (00:54:06) Regulation and Geopolitical Landscape in AI (00:55:58) Foundation Models and Frontier Models (01:01:06) Economics and Open Source Models (01:05:18) AI and AGI Accessibility (01:07:42) Optimizing AI Prompts (01:11:18) Creating an AI Superpower (01:14:12) Future of Education and AI (01:19:34) The Impact of AI on Job Roles (01:21:58) AI in Problem-Solving and Research (01:25:24) Importance of AI Context Window (01:27:37) AI Output and Intellectual Property (01:30:09) Google Maps Launch and Challenges (01:37:57) Long-Term Investment in AI (01:43:02) Balancing Work and Family Life (01:44:25) Building Sierra as an Enduring Company (01:45:38) Lessons from Tech Company Lifecycles (01:48:31) Definition and Applications of AI Agents (01:53:56) Challenges and Importance of Branded AI Agents (01:56:28) Fending Off Complacency in Companies (02:01:21) Customer Obsession and Leadership in Companies Bret Taylor is currently the Chairman of OpenAI and CEO of Sierra. Previously, he was the CTO of Facebook, Chairman of the board for X, and the Co-CEO of Salesforce. Newsletter - The Brain Food newsletter delivers actionable insights and thoughtful ideas every Sunday. It takes 5 minutes to read, and it’s completely free. Learn more and sign up at fs.blog/newsletter Upgrade — If you want to hear my thoughts and reflections at the end of the episode, join our membership: fs.blog/membership and get your own private feed. Watch on YouTube: @tkppodcast Learn more about your ad choices. Visit megaphone.fm/adchoices
Transcript
Discussion (0)
Technology companies aren't entitled to their future success.
AI, I think, will change the landscape of software,
and I think it will help some companies, and it will really hurt others.
And so when I think about what it means to build a company that's enduring,
that is a really, really tall task in my mind right now,
because it means not only making something that's financially enduring over the next 10 years,
but setting up a culture where a company can actually evolve to meet
the changing demands of society and technology when it's changing at a pace that is like
unprecedented history. So I think it's one of the most fun business challenges of all time.
I just get so much energy because it's incredibly hard and it's harder now that it's ever been
to do something that lasts beyond you. But that I think is the ultimate measure of a company.
podcast. I'm your host, Shane Parrish. In a world where knowledge is power, this podcast is your
toolkit for mastering the best of what other people have already figured out.
If you want to take your learning the next level, consider joining our membership program at
fs.blog slash membership. As a member, you'll get my personal reflections at the end of
every episode, early access to episodes, no ads, including this, exclusive content, hand-edited
transcripts and so much more. Check out the link in the show notes for more. Six months after
Brett Taylor realized AI was about to change everything, he walked away from his co-CEO job at
Salesforce to start from scratch. That's how massive this shift really is. The mastermind behind
Google Maps and the former chief technology officer at Facebook, Brett reveals the brutal
truce about leadership, AI and what it really takes to build something that endures long after
you've reached the top. Brett's led some of the most influential companies in tech and seen exactly
what makes businesses scale, what kills them from within, and why most founders don't survive
their own success. In this conversation, you'll discover why so many companies are already on life
support without realizing it, how first principles thinking separates the next wave of winners from
everyone else and the hidden reason most acquisitions fail. We'll explore why AI is bigger than anyone suspects,
plus the mindset shift that turns great engineers into exceptional CEOs.
Whether you're a founder, an operator, or simply someone who wants to think sharper,
this episode will change how you see your business, technology, and the future.
It's time to listen and learn.
What was your first real aha moment with AI where you realized, holy shit, this is going to be huge?
I had two separate aha moments, one that I don't think I really appreciated how huge it would be,
but it kind of reset my expectation, which was the launch of Dolly in the summer of 22, is that right?
That might be off a year, but I think summer of 22, and the avocado chair that they generated.
And I had been, well, my background is in computer science and pretty technically deep.
I hadn't been paying attention to large language models.
I just didn't follow the progress after the Transformers paper.
And I saw that and my reaction was I had no idea computers could do that.
And that particular launch, you know, seeing a generated image of an avocado chair,
I don't think I extrapolated to what, you know, where we are now.
But it, for me, shook me and realized I'd need to pay more.
attention to this space and Open AI specifically than I had been.
I think I had, that was the moment where I realized, like, I clearly have been not paying attention
to something significant.
And then it was, you know, six months later, coincidentally, like the month after I left
Salesforce, the chat CBD came out and before it became a phenomenon, though it did so
quickly, but I was already, you know, plugged into it.
And I was, from that on, you know, I could not stop thinking about it.
But that avocado chair, I don't know why.
I think it was the, there was a bit of an emotional moment where you saw a computer doing something that wasn't just rule-based, but creative.
And the idea of a computer doing something, creating something from scratch was, well, it doesn't seem so novel.
You know, a few years later, just blew my mind at the time.
One of the unique things about you is that you've started companies, you've been acquired by Facebook and Salesforce.
Inside those companies, you rose up to be the CTO at Facebook, the co-CEO at Salesforce.
Talk to me about founders working for founders and founders working within a company.
Yeah, it's a very challenging transition for a lot of founders to make.
I think there's lots of examples of acquisitions that have been really transformative from a business standpoint.
I think YouTube, Instagram, being two of the more prominent that have clearly changed the shape of the acquiring company.
But even in those cases, you know, the founders didn't stay around that long.
And that's maybe a little unfair.
Stick around for a little bit.
I think the interesting thing about being a founder is it's not just building a business, but it's very much your identity.
And I think it's very hard for people who aren't founders to experience it.
You take everything very personally, you know, from the product to the customers, to the press, to your competitors, the both inner and outer measures of success.
And I think when you go to being acquired, there's a business aspect to it.
And can you operate within a larger company?
But that's intertwined with a sense of identity.
You go from being the founder of a company and the CEO of a company or CTO of a company, whatever your title happens to be, as one of the co-founders.
to be in a part of a larger organization and to fully embrace that, you actually need to change
your identity. You need to go from being, you know, the head of Instagram, in my case, the head
of QIP to being an employee of Salesforce or going from being the CEO of Friend Feed to being an
employee of Facebook. And what I've observed is it's that identity shift is a prerequisite for most
of the other things. It's not simply your ability to handle the politics and bureaucracy of a bigger
company or to navigate a new structure. I actually think most founders don't make that leap
where they actually identify with that new thing. It's even harder for some of the employees too
because most of the time in an acquisition, an employee of an acquired company didn't choose that
path. And in fact, they chose to work for a different company and they, you know, the acquisition
determined a different outcome. And that's why integrating acquisitions is so nuanced. And I would
say that having the experience of having been acquired, you know, before and having acquired
some companies before when I got to Salesforce, I really tried to be self-aware about that and really
tried to, you know, be a part of Salesforce, you know, and tried to shift my identity and
not be a single-issue voter around QIP. You know, I'd really try to embrace it. And I think
it's really hard for some founders do. And some founders don't want to, honestly, you know, they
maybe cash the check and and you know that's the it's more of a transactional relationship I
I really actually am so grateful for the experience of having been at Facebook and Salesforce I
learned so much so but it really took a lot of effort on my part to just um transform my perception
of myself and who I am to get that value out of the company that acquired us how did you how did
it change how you did acquisitions at Salesforce you guys did a lot of acquisitions what you were there
and you're acquiring founders and sort of startups,
and I think Slack was while you were there too.
How did that change how you went about integrating that company
into the Salesforce culture?
I'll talk abstract about talking about some specific acquisitions too,
but first I think I tried to approach it with more empathy
and more realism.
You know, one of the nuanced parts about acquisitions is
there's the period of doing the acquisition,
there's sort of the period after you've decided to do it of doing due diligence,
and then there's a period when it's done,
and you're integrating the company and sort of the period after.
One of the things that I have observed is that companies doing acquisitions,
often the part of deciding to do it is a bit of a mutual sales process.
You're trying to find a fair value for the company,
and there's some back and forth there.
But at the end of the day, there's usually some objective measure
of that influenced by a lot of factors, but there's some fair value of that. But what you're trying
to do is what are, and corporate speak, would be synergies, but like, why do this? Why is
one plus one greater than two? You know, that's why you do an acquisition just from first principles.
It's often an exercise in storytelling. You know, you bring this product together with our product,
and customers will, you know, find the whole greater than the sum of its parts. This team applies,
to our sales channel or if you're a Google acquisition, you know, imagine the traffic we can
drive to this product experience, you know, in the case of something like an Instagram,
imagine our ad sales team attached to your amazing product and how quickly we can help
you realize that value, whatever it might be. I find that people, because there's sort of
a craft of storytelling, for both sides to come to the same conclusion that they should do
this acquisition, sometimes either simplifies or sugarcoats like some of the realities of
it, you know, little things like, you know, how much control will the founding team of the
acquired company have over those decisions? Will it be operated as a standalone business
unit or will your team be sort of broken up into functional groups within the larger company?
and it's sort of those little, they're not little, but those I'll say boring but important things
that often people don't talk enough about. And you don't need to figure out every part of an
acquisition to make it successful, but often you can end up running into like true third rails
that you didn't find because you were having the storytelling discussions rather than getting
down to brass tax about how things will work and what's important. The other thing that I think is
really important is being really clear what success looks like. And, you know, I
I think sometimes it's a business outcome, sometimes it's a product goal, but I found that
if you went to most of the larger acquisitions in the Valley, and you, two weeks after it was
closed, interviewed the management team of the acquiring company and the acquired company,
and you asked them, like, what does success look like two years from now?
My guess is, like, 80% of the time you get different answers.
And I think it goes back to this sort of storytelling thing where you're talking about the benefits of the acquisition about talking about like, what does success look like?
So I really tried to approach it.
I tried to pull forward some harder conversations when I'm doing acquisitions or even when I'm being acquired since it's happened to be not twice.
So that, you know, when you're approaching it, you'd not only get the, hey, why is one plus one equal gradient to two?
Everything's going to be awesome, you know.
but no for real like what you know what does success look like here and then you know as a founder
your job of an acquired company is to tell your team that and align your team to that and I think
founders don't take on enough accountability towards making these acquisitions successful as I think
they should and and it goes back to again a certain uh naivete you know it's like you're you're
not your company anymore you're part of something larger and I think you know successful ones work
when everyone embraces, embraces that.
What point in the acquisition process is that conversation?
Is that after we've signed our binding, you know, sort of commitment?
Or is it we should have that conversation before so I know what I'm walking into?
My personal take is it's not something you have, you have to get to the point where the two parties want to merge, you know.
And that's obviously a financial decision, particularly if it's like a public company.
there's a board and shareholders.
Most acquisitions in the Valley are a larger firm acquiring a private firm.
That's not all of them, but I would say that's the vast majority.
And in those cases, there's often a qualitative threshold where someone's like, yeah, let's do this.
We kind of have the high-level terms, sometimes a term sheet, you know, formally.
I think it's right after that.
So where people have really committed to the key things, how much value, why are we doing this, the big stuff.
and there's usually, you know, many, lots of lawyers being paid lots of money to turn
those term sheets into, you know, more complete set of documents, usually more complete
due diligence, stuff like that.
That's a, there's an awkward waiting period there.
And that's a time, I think, where, like, the strategic decision makers in those
moments can get together and say, let's talk through what this really means.
And the nice part about having them for all parties is you've kind of made the commitment to
each other so it's you've I think you have more social permission to have real conversations at that
point but you also haven't consummated the relationship you know and so there's a the power imbalance
isn't totally there and and you can really talk through it and it also I think engender's trust
just because by having a harder conversations in those moments you're learning how to have real
conversations and learning how each other works so that's my my personal opinion when to have it
So you mentioned the board, you've been on the board of Shopify, you're on the board of OpenAI, you're a founder.
What's the role of a board and how is it different when you're on the board of a founder-led company?
With Amex Platinum, $400 in annual credits for travel and dining means you not only satisfy your travel bug, but your taste buds too.
That's the powerful backing of Amex. Conditions apply.
Hit pause on whatever you're listening.
to and hit play on your next adventure.
This fall get double points on every qualified stay.
Life's the trip.
Make the most of it at BestWestern.
Visit bestwestern.com for complete terms and conditions.
I really like being involved in a board, and I've been involved in multiple boards
because I think I am an operator through and through.
I probably self-identifies an engineer first more than anything else, and I love to build.
learning how to be an advisor is a very different vantage point that I think you see how other
companies operate and you also learn how to have an impact and add value without doing it
yourself and it's a very and I've really I think become a better leader you know having learned
to do that I have really only joined boards that were led by founders because typically I think
they, you can speak to them, but I think they sought me out because I'm a founder and I like
working with founder-led companies. I think the founders, I'm sure there's lots of studies on this,
but I think founders drive better outcomes for companies. There's a, I think founders tend to have
permission to make bolder, more disruptive decisions about their business than a professional
manager. There's exceptions like Sotcha, I think, is, you know, one of the greatest,
if not the greatest CEO of, you know, our generation and as a professional manager. But, you know,
you look at everyone from Toby Lukie to Mark Benioff to Mark Zuckerberg to Sam at OpenAI.
And I think when you have founded a company, all your stakeholders, employees in particular give
you the benefit of the doubt. You know, you created this thing. And if you say, hey, we need to
do a major shift in our strategy, even hard things like layoffs. Founders tend to get a lot of
latitude and are judged, I think, differently. And I think rightfully so in some ways because of the
interconnection of their identity to the thing that they've created. And so I actually really
believe in founder-led companies. One of the real interesting challenges is going from a founder-led
company to not. And, you know, Amazon has gone through that transition. Microsoft has gone
through that transition for that reason. But I love working with founders. And I, I love working
with people like Toby and Sam because they're so different than me yet. And I can see how they
operate their businesses and I am inspired by it. I learn from it. And obviously, working for market.
Salesforce, you, I'm like, wow, that's really interesting. I'm almost like an anthropologist.
Why did you do that? You know, I want to learn more. And so I love working with founders that inspire me
because I just learned so much from them.
It's such an interesting front row seat into what's happening.
Do you think founders go astray when they start listening to too many outside voices?
And this goes back to the, I'm sure you're aware, the Brian Chesky, the founder mode.
Do you think, talk to me about that.
I have such a nuanced point of view on this because it is decidedly not simple.
So broadly speaking, I really like the spirit of founder mode, which is just having.
having deep founder-led accountability for every decision at your company. I think that that's how great companies operate. And when you, you know, proverbially make decisions by committee or you're more focused on process than outcomes, that produces all the experiences we hate as employees, as customers. That's the proverbial DMV, right? You know, it's like process over outcomes. And then similarly, you look at the,
disruption in all industries right now because of AI, you know, the companies that will
recognize where things are clearly going to change. Like, everyone can see it. It's like a,
you know, slow motion car wreck. Everyone knows how it ends. You need that kind of decisive
breakthrough, boundaries, layers of management to actually make change as fast as required
in business right now. The issue, I have not with Brian's statements, Brian's amazing,
is how people can sort of interpret that
and sort of execute it as a caricature
of what I think it means.
You know, there was a,
I remember after Steve Jobs passed away
and, you know, I don't know,
I've met Steve a couple times
I haven't never worked with him
in any meaningful way.
You know, but he was sort of,
if you believe the story is like kind of
pretty hard on his employees
and very exacting.
And I think a lot of founders were like mimicking that,
you know, done to wearing a black turtlene,
and yelling at their employees.
I'm like, I'm not sure that was the cause.
You know, I think Steve Jobs' taste and judgment
through, you know, executed through that, you know,
packaging was the cause of their success.
And somehow, and then similarly, I think founder mode can be weaponized
as an excuse for just like overt micromanagement.
And that probably won't lead to great outcomes either.
And most great companies are filled with extremely great individual contributors
who make good decisions and work really hard.
and companies that are like solely executing through the judgment of individual probably aren't
going to be able to scale to be truly great companies. So I have a very nuanced point because I actually
believe in founders. I believe in actually that accountability that comes from the top. I believe
in cultures where, you know, founders have license to go in and all the way to a small decision
and fix it, the infamous question mark emails from Jeff Bezos, you know, that type of thing.
That's the right way to run a company.
But that doesn't mean that you don't have a culture where individuals are accountable and empowered.
And you don't want people trying to make business decisions because of what will please our dual leader, which is like the caricature of this.
And so after that came out, I could sort of see it all happening, which is like some people will take that.
I'm like, you know what?
You're right.
I need to go down and be in the details.
And some people will do it and probably make everyone who works for them miserable and probably both will happen as a consequence.
totally thank you for the detail and nuance there i love that too do you think engineers make good
leaders i do think engineers make good leaders but one thing i've seen is that i think that i really
believe that great CEOs and great founders um start usually with one specialty but become
uh more broadly specialists in our parts of their business um you know i think
I think the businesses are multifaceted and rarely as a business's success due to one thing, like engineering or product, which is where a lot of founders come from.
Often your go-to-market model is important for consumer companies.
How you engage with the world in public policy becomes extremely important.
And I think as you see founders grow from doing one thing to growing to being a real meaningful company like Airbnb or meta or something,
you can see those founders really transform from being one thing to many things.
So I do think engineers make great leaders.
I think the first principles thinking, the system design thinking, really benefits things like organization design, strategy.
But I also think that, you know, when we were speaking earlier about identity, I think one of the main transitions founders need to make, especially engineers,
is you're not like the product manager for the company, you're the CEO.
And at any given day, do you spend time recruiting an executive because you have a need?
Do you spend time on sales because that will have the biggest impact?
Do you spend time on public policy or regulation?
Because if you don't, it will happen to you and could really impact your business in a negative way.
And I think engineers who are unwilling to elevate their identity from what they were to what it needs to be in the moment often leads to sort of plateaus in companies growth.
So 100 percent, I think engineers make great leaders.
And it's not a coincidence, I think, that most of the Silicon Valley, great Silicon Valley CEOs came from engineering backgrounds.
But I also don't think that's sufficient either as your company scales.
And I think that making that transition, as all the great ones have, is incredibly important.
To what extent are all business problems, engineering problems?
That's a deeper philosophical question that I think I have the capacity to answer.
What is engineering?
What I like about approaching problems as an engineer is first principles thinking and understanding the root causes of issues rather than simply addressing the symptoms of the problems.
And I do think that coming from a background in engineering, that is everything from process, like how engineers do a root cause analysis of an outage on a server is a really great way to analyze why you lost a sales deal.
You know, like I love the systematic approach of engineering.
One thing that I think going back to good ideas that can become caricatures of themselves, like one thing I've seen though with engineers who go into other disciplines is sometimes you can over.
overanalyze decisions in some domains. Let's just take modern communications, which is driven
in social media and very fast-paced. Having a systematic first principles discussion about
every tweet you do is probably not a great comm strategy. And so, and then similarly, you know,
there are some aspects of, say, enterprise software sales that, you know, aren't rational, but
they're human, you know, like forming personal relationships, you know, and the importance of those
to building trust with a partner. It's not all just, you know, product and technology. And so I would
say I think a lot of things coming with an engineer mindset could really benefit, but I do think
that taking that to its, like, logical extreme can lead to analysis paralysis, can lead to
over-intellectualizing some things that are fundamentally human problems. And so
So, yeah, I think a lot can benefit from engineering, but I wouldn't say everything's an engineering problem in my experience.
You brought up first principles a couple of times. You're running your third startup now, Sierra. It's going really well. How do you use first principles in terms of how do you use that at work?
Yeah, it's particularly important right now because the market of AI is changing so rapidly. So if you rewind two years, you know, most people,
hadn't used chat GPT yet. Most companies hadn't heard the phrase large language models or
generative AI yet. And in two years, you have chat GPT becoming one of the most popular
consumer services in history faster than any service in history. And you have across so many
domains in the enterprise, really rapid transformation. The law is being transformed.
marketing is being transformed, customer service, which is where my company,
Sierra Works is being transformed. Software engineering is being transformed.
And the amount of change in such a short period of time is, I think, unprecedented.
And perhaps I lack the historical context, but it feels faster than anything I've experienced in my career.
And so as a consequence, I think, if you are responding to the facts in front of you
and not thinking from first principles about why we're at this point
and where it will probably be 12 months from now,
the likelihood that you'll make the right strategic decision is almost zero.
So as an example, it's really interesting to me
that with modern large language models,
one of the careers that is being most transformed is software engineering.
And, you know, one of the things I think a lot about is
how many software engineers will we have our company three years from now? What will the role
of a software engineer be as we go from being authors of code to operators of code generating
machines? What does that mean for the type of people we should recruit? And if I look at the actual
craft of software engineering that we're doing right now, I think it's literally a fact that
will be completely different to years from now. Yet I think a lot of people building companies
hire for the problem in front of them rather than doing that.
But two years is not that long.
Those people that you hire now will just be getting really productive a couple years from now.
So we try to think about most of our long-term business from First Principles, everything from, I'll say a couple examples in our business.
Our pricing model is really unique and comes from First Principles thinking rather than having our customers pay a license for the privilege of using our platform.
we only charge our customers for the outcomes, meaning if the AI agent they've built for their
customers solves the problem, there's usually a pre-negotiated rate for that.
And that comes from the principle that in the age of AI, software isn't just helping you be
more productive, but actually completing a task.
What is the right and logical business model for something that completes a task?
Well, charging for a job well done, rather than charging for the privileges using the software.
Similarly, you know, we with a lot of our customers, you know, we help deliver them a fully working AI agent.
We don't hand them a bunch of software and say, good luck, you know, configure it yourself.
And the logic there is, you know, in a world where making software is easier than it ever is before
and you're delivering outcomes for your customer, the delivery model of software probably should change as well.
And we've really tried to reimagine what, like, the software company of the future should look like
and trying to, you know, model that in everything that we do.
That's brilliant.
How do you think software engineering will change?
Is that you're going to have fewer people or the people are going to be organized differently?
Or how do you see that?
How geeky can I get?
As geeky as you want, man.
I actually wrote a blog post right before Christmas about this.
I think this is an area that deserves a lot more research.
I'll describe where I think we are today, and smart people may disagree, but a lot of the modern large language models, both the traditional large language models and sort of the new reasoning models are trained on a lot of source code, and it's an important input to all of the knowledge that they're trained on.
As a consequence, even the early models were very good at generating code.
So, you know, every single engineer at Sira uses a cursor, which is a great product that basically integrates with the IDE Visual Studio code to help you generate code more quickly.
It feels like a local maximum in a really obvious way to me, which is you have a bunch of code written by people, written in programming languages that were designed to make it easy for people.
to tell a computer what to do.
Probably this funniest example of this is Python.
It almost looks like natural language.
Yeah.
But it's notoriously not robust.
You know, most Python bugs are found by running the program
because there's not static type checking.
Similarly, there's most bugs, while you could run a fancy static analysis,
like most bugs show up simply at runtime because it's just not designed.
It's designed to be ergonomic to write.
Yet, we're using AI to generate that.
And so we've sort of designed most of our computer programming systems to make it easy for the author of code to type it quickly.
And we're in a world, we're actually generating code is going to, like, the marginal cost of doing that is going to zero.
But we're still generating code in programming languages that were designed for human authors.
And similarly, if you've ever looked at someone else's code, which a lot of people do professionally
is called a code review, it's actually quite hard to do a code review.
You know, you end up interpreting, you're trying to basically put the system in your head
and simulate it as you're reading the code to find errors in it.
So the irony now that have taken things that are code programming languages that were designed for authors,
And now having humans do the job of essentially code reviewing,
code written by an AI,
and yet all of the AI being in the code generation part of it,
I'm like, I'm not sure it's great,
but we're generating a lot of code with similar flaws
to that we've been generating before from security holes
to just functional bugs and in greater volumes.
And I think what I would like to see is
if you start with the premise that generating code is free
or going towards free, what would be the programming systems that we would design?
So, for example, Rust is an example of a programming language that was designed for safety,
not for programming convenience.
You know, my understanding is that the Mozilla project, you know, there were so many security
holes in Firefox.
They said, let's make a programming language that's very fast, you know, but everything can
be checked statically, including memory safety.
Well, it's a really interesting direction where you weren't optimizing for authorship convenience or optimizing for correctness.
Are there programming language designs that are designed so a human looking at it can very quickly evaluate, does this do what I intended it to do?
There's an area of computer science I studied in college called formal verification, which at the time was turning a lot of computer programs into math proofs and finding inconsistencies, and it sort of worked well as not as well as you'd hope.
But, you know, in a world where AI is generating a lot of code, you know, should we be investing more informal verification so that the operator of that code generator machine can more easily verify that it does, in fact, to do what they intended is to do, and could a combination of a programming language that is more structurally correct and structurally safe and exposes more primitives for verification plus a tool to verify, could you make an operator of a code generating machine?
20 times more productive, but more importantly, make the robustness of their output 20 times greater.
And then similarly, you know, there's themes, things go in and out of fashion, but like test
driven development, you know, where you write your unit test first or your integration test
first and then write code until it fulfills the test. Most programmers I know who are really good,
not despise it, but it's just like a, it sounds better than it is in practice. But again,
writing code is free, you know, so writing tests is free. You know, how can you create a programming
system where the combination of great programming language design, formal verification,
robust tests because you didn't have to do the tedious part of writing them all, could you make
something that made it possible to write increasingly complex systems that were increasingly
robust? And then similarly like the elephant in the room for me is the anchor tenant of most of these
code generating systems are an IDE right now, you know, and that obviously doesn't seem as
important in this world. And even with coding agents, which is sort of where the world is going,
it doesn't change the fact that, like, you know, who's accountable for the quality of it,
who's fixing it. And I think there is a world where we can make reasonable software by just
automating what we as software engineers do every day. But I have a strong suspicion that
if we designed these systems with the role of a software engineer in mind being an operator
of a machine rather than the author of the code, we could make the process much more robust and
much more productive. And it feels like a research problem to me. It doesn't feel, and I think a lot
of people, and for a good reason, including me, are just excited about the efficiency of software
development going up. And I want to see the new thing. I'm constructively dissatisfied with where we
are. It's so interesting that if software AI is good enough to write the code, should we get
enough check the code? That's a great question. But actually,
you'll, you know, it's still funny to me that we'd be generating Python, you know, just because
for anyone who's listening right now has ever operated a web service running Python, it's CPU
and intensive, really inefficient, you know, should we be taking most of the unsafe C code that
we've written and converting it to a safer system like Rust? You know, if authoring these things
and checking it are relatively free, shouldn't all of our programs be incredibly efficient?
Should they all be formally verified? Should they all be?
analyzed by a great agent. I do think it can be turtles all the way down. You can use AI to
solve most problems in AI. The thing that I'm trying to figure out is like what is the system
that a human operator is using to orchestrate all those tasks? And, you know, I go back to the
history of software development. And most of the really interesting metaphors in software
development came from breakthroughs in computing. So, you know, the C programming language came
from Unix and when these time sharing systems were really, it went from sort of punch cards to
something that were a lot more agile. Small talk came out of the development of the graphical
user interface at Xerox Park and, you know, there was a sort of a confluence of message passing
as a metaphor and the graphical user interface. And then there was a lot of really interesting
principles that came out of networking, you know, and sort of distributed systems, distributed
locking, sequencing, I think we should recognize that we're in this brand new era as
significant as the GUI. It's like a completely new era of software development. And if you were
just to say, I'm going to design a programming system for this new world from first principles,
what would it be? And I think when we develop it, I think it will be really exciting because
rather than automating and turning up the speed of just generating code and with the same
processes we have today, I think will feel native to this system.
and give a lot more control to the people who are orchestrating the system in a way that I think will really benefit software overall.
Let's dive into AI a little bit. How would you define AGI to the layman?
I think a reasonable definition of AGI might be that any task that a person can do at a computer, that system can do on par or better.
I'm not sure that's a precise definition, but I'll tell you where that comes from and it's flaws,
but there's not a perfect definition of AGI, in my opinion, or there's not a precise definition of AGI.
I'm sure there's good answers.
One of the things about the G and AGI is about generalization.
So can you have a system that is intelligent in domains that it wasn't explicitly trained to be intelligent on?
And so I think that's one of the most important things is like given a net new domain, can this system become more competent and more intelligent than a person sort of trained in that domain?
And I think that's sort of the, you know, at or better than a person is certainly a good standard there.
And that's sort of the definition of superintelligence.
the reason I mentioned at a computer is I do think that it is a bar that means
like if there's a digital interface to that system, it affords the ability for AI to interact
with it, which is why that's a bar that's reasonable to hit.
I say that because one of the interesting questions around AGI is how quickly it does
generalize. And there are domains in the world that the progress in that domain isn't necessarily
limited by intelligence, but by other social artifacts. So as an example, and I'm not an expert
in this area, but if you think about the pharmaceutical industry, my understanding is, you know,
the one of the main
bottlenecks
is clinical trials
so no matter how
intelligent a system would be
in discovering new
therapies
it may not
materially change that
and so you may have something that's
discovering new insights
in math and that would be
delightful and amazing
but the existence
of that system
that's super intelligent in one domain
may not translate to all domains equally.
I just heard at least a snippet of a talk by Taller Cohen, the economist,
and it was really interesting to hear his framing on this
about which parts of the economy could sort of absorb intelligence more quickly than others.
And so I choose that definition of AGI,
recognizing that there's not a perfect definition
because it captures the ability of this intelligence to generalize
while also recognizing that the domains of society
like it might not apply with equal velocity
even once we reach that point of a system
being able to have that level of intelligence.
You can get protein at home
or a protein latte at Tim's.
No powders, no blenders, no shakers.
Starting at 17 grams per medium latte,
Tim's new protein lattes,
protein without all the work,
at participating restaurants in Canada.
Summer's here and you can now get
almost anything you need for your sunny day.
delivered with Uber Eats.
What do we mean by almost?
Well, you can't get a well-groom lawn delivered,
but you can get a chicken parmesan delivered.
A cabana? That's a no, but a banana, that's a yes.
A nice tan, sorry, nope.
But a box fan, happily yes.
A day of sunshine? No.
A box of fine wines?
Yes. Uber Eats can definitely get you that.
Get almost, almost anything delivered with Uber Eats.
Order now.
Alcohol and select markets.
Product availability may vary by Regency app for details.
When I think about what
artificial intelligence
is limited by
or the bottlenecks if you will
I keep coming back to a couple things
there's regulation
there's compute
there's energy
there's data
and there's LMs
am I missing anything
so you're saying
the ingredients to AGI
yeah like there's limitations
on each aspect of those things
and there seem to be the main contributors
to what's limiting us
from even accelerating
at this point is that
how do you think
about that? Yeah, what you said is roughly how I think about it. I'll put it into my own words,
though. I think the three primary inputs are data, compute, and algorithms. And data is probably
obvious, but one of the things after the Transformer model was introduced is it afforded an architecture
with just much greater parallelism, which meant models could be much bigger and train more quickly
on much more data, which just led to a lot of the breakthroughs with, that's the LLM, just they're
large.
Yeah.
And the scaling laws, you know, a couple of years ago, you know, indicated, like, the larger
you make the model, the more intelligent would be at a degree of efficiency that was tolerable.
And there, we are, you know, there's lots of stuff written about this, but, you know, there's, in terms of just, like, text.
content to train on, you know, the availability of new content is certainly waning. And some
people would say I think there's like a data wall. I'm not an expert in that domain, but it's
been talked about a lot and you can read a lot about it. There's a lot of interesting opportunities
though to generate data too. So there's a lot of people working on simulation. If you think about
a domain like self-driving cars, simulation is a really interesting way to generate.
Is that synthetic data? Is that what? Yeah, I would say that synthetic data. The synthetic data
has a simulation and synthetic data are a little different.
So you can generate synthetic data, like generate a novel.
Simulation, I would put at least in my head, and I'm sure that academics might critique what I'm saying.
But I view simulation is based on a set of principles like the laws of physics.
So if you're able to build a real world simulation for training a self-driving car,
you're not just generating arbitrary data.
like the roads don't turn into loop-de-loops, you know, because that's not possible with physics.
So by constraining a simulation with a set of real-world constraints, the data has more efficacy, you know?
And so, and there's sort of a, it constrains the different permutations of data you can generate from it.
So it's, I think, a little bit higher quality.
But then along those lines, you know, a lot of people wonder if you generate synthetic data,
How much value can that add to a training process?
You know, is it sort of regurgitating information already had?
What's really interesting about, you know, reasoning and reasoning models is I think I feel
really optimistic these models are generating net new ideas, and so it really affords
the opportunity to break through some of these, the data wall as well.
So data is one thing, and I think both synthetic data and simulation are really interesting opportunities
to grow there.
Then you have compute.
And this is something that, you know,
that's why there's so many data center investments.
It's why NVIDIA as a company has grown so much.
Probably the more interesting kind of breakthroughs there
are these reasoning models where there's not quite such a formal separation
between the training process and the inference process
where you can spend more compute at,
the time of inference to generate more intelligence,
which has really been a breakthrough in a variety of ways,
I think is really interesting,
but it shows you how you can run up against walls
and find new opportunities to use it.
And then finally, algorithms.
And the biggest breakthrough is obviously the Transformers model,
attention is all you need, that paper from Google,
that sort of led to where we are now.
But there's been a number of really important papers
since then from the idea of chain of thought reasoning
to what at Open Eye, what we did with the O1 model,
which is to do some reinforcement
or in those chains of thought
to really reach new levels of intelligence.
And so I do think that
I mentioned some anecdotes
about some breakthroughs there
because my view is that each one of them
has their own problems.
You know, compute.
It's very capital intensive.
And a lot of these models,
the half-life of their value is pretty short
because new ones come out so frequently.
And so, you know, you wonder,
like, you know, can we afford? What's the business case for investing this CapEx? And then you
have a breakthrough like, you know, oh one and you're like, gosh, you know, with a distilled
model and moving more to inference time, it changes the economics of it. You have data. You say,
gosh, we're running out of textual data to train on. Well, now we can generate reasoning.
We can do simulations. Oh, that's an interesting breakthrough. And then on the algorithm side,
as I mentioned, just the idea of these reasoning models is really novel itself. And each of these
at any given point, if you talk to an expert in one of them and I'm an expert in none of them,
they will tell you the sort of current plateau that they can see on the horizon. And there usually
is one. I mean, you'll talk to different people about how long the scaling laws for something
will continue and you'll get slightly different opinions, but no one thinks it's going to last forever.
And at each one of those, because you have so many smart people working on them,
you often have people discovering a breakthrough in each of them. And so as a constant,
I really do feel optimistic about the progress towards H.E.I. Because one of those plateaus might
extend a while if we just don't have the key idea that we need to break through. The idea that
we will be stuck on all three of those domains feels very unlikely to me. And in fact,
what we've seen because of the potential economic benefits of AGIs, we're in fact seeing
breakthroughs in all three of them. And as a consequence, you know, you're just seeing just the
blistering pace of progress that we've seen over the past couple years.
At what point does AI start making AI better than we can make it, or making it better while
we're sleeping? We can't be too far from that. Well, it might reflect back to our software engineering
discussion, but broadly, this is the area of AGI around self-improvement, which is meaningful from
an improvement standpoint, but also obviously from a safety standpoint as well. So,
I don't know when that will happen, but I do think, you know, by some definition, you could argue that it's happening already in the sense that every engineer in Silicon Valley is already using coding agents and platforms like Cursor to help them code. So it's contributing already.
And I imagine as coding assistants go to coding agents in the future, most engineers in Silicon Valley will show up in the morning.
But this is sort of the different.
between, you know, the assisted driving and Tesla versus, like, self-driving.
Yeah.
Right?
Like, at what point do we leap from, I'm a co-pilot in this to, I don't have to do anything?
I mean, it's a question that there's so much nuance to answer.
I'm not sure to answer because I'm not sure you'd want to necessarily.
Like, I think for some software applications, that's important.
But when we brought up, you know, we were talking about the active software development,
people have to be accountable for the software that they produce.
And that means if you're doing something simple like a software-as-a-service application,
that it's secure, that it's reliable, that the functionality works as intended
for something as meaningful as an agent that is somewhat autonomous,
does it have the appropriate guardrails, does it actually do what the operator's intended,
is there appropriate safety measures.
So I'm not sure there's really any system
where you'd want to turn a switch
and go get your coffee.
But I do think to the point on, you know,
these broader safety things
is I think that when you think about more advanced models,
we need to be developing not only more and more advanced
safety measures and safety harnesses,
but also using AI to supervise AI and things like that.
So it's a part, probably my colleague
on the board Zico Coulter is probably a better person to talk through some of the technical
things, but there's a lot of prerequisites to get at that point, and I'm not sure it's simply
like the availability of the technology, just because it is that at the end of the day,
we are accountable for the safety of the systems we produce, not just opening out, like every
engineer, and that's a principle that should not change.
What does that mean, like when we say safety and AI, that seems so vague in general that
everybody interprets it quite differently.
Like, how do you think about that?
And how do you think about that in the world where, let's say, we regulate safety in
the United States and another country doesn't regulate safety?
How does that affect the dynamic of it?
I'll answer broadly and then go into the regulatory question.
So I really like Open AI's mission, which is to ensure that AGI benefits all of humanity.
that isn't only about safety
and I believe intentionally so
though obviously the mission
was created prior to my arrival
because it's both about safety
kind of hypocritic of first do no harm
and I don't think one could credibly achieve that mission
if we created something unsafe
so I would say that's the most important part of the mission
but there's also a lot of other aspects
of benefiting humanity
is it universally accessible
is there a digital divide where some people have access
to AGI and some day
don't. Similarly, you could argue that does it, are we maximizing the benefits and minimizing
the downsides? Clearly, AI will disrupt some job, but it also could democratize access to
health care, education, expertise. So as I think about the mission, it starts with safety,
but I actually like thinking about it more broadly because I think at the end of the day,
benefiting humanity is the mission and safety is a prerequisite. But, safety is a prerequisite, but
it's almost like going to my analogy of the Hippocratic Oath, a doctor's job is, you know, to cure you.
First, do no harm, but then to cure you.
And a doctor that did no harm but didn't cure you wouldn't be great either.
So I really like to think about the holistically.
And again, Zika or Sam might have a more complete answer here.
But broadly, I think about does the system that represents AGI align with the intentions of the people created it and the intentions that people operate?
it so that it does what we want and it's a tool that benefits humanity, a tool that we're
actively using to affect the outcomes that we're looking for. And that's kind of the way I think
about safety. And it can be meaningful things like misalignment or more subtle things like
unintended consequences. And I think that latter part is probably the area that's really
interesting from an intellectual and ethical standpoint as well.
If I look at what was the bridge in Canada that fell down where it motivated the ring that a lot of engineers were.
Oh, yeah, I forget the name of it.
Just like whether it's the Tacoma Narrows Bridge in Washington or Three Mile Island or these intersections where we've engineered these, you know, what at the time people hope would be positively impact humanity, but something went horribly wrong.
Sometimes it's engineering, sometimes it's bureaucracy, sometimes it's a lot of things.
And so I don't think, when I think about safety, I don't just look at the technical measures of it, but how does this technology manifest in society?
How do we make decisions around it?
And you could take, put another way, technology is rarely innately good or bad.
It's sort of what we do with it.
And I think those social constructs matter a lot as well.
So I think it's a little early to tell because we don't have this kind of superintelligence right now.
And I think it won't just be a technology company defining how it manifests in society.
And you could imagine taking a very well-aligned AI system and a human operator directing it towards something that would objectively hurt society.
And there's a question of like, who gets to decide, who's accountable?
And it's a perennial question.
I mean, it's whether you're deciding, you know, should you use your.
smartphone in school, you know, who should decide that. And I, there's parents who will tell
you, hey, it's my decision. It's my kid. And then there's principals who will tell you it's not
benefiting the school. And I'm not sure that's going to be my place or our place, but there'll be
a number of those conversations that are much deeper than that question that I think we'll need to
answer. As it relates to regulation, there's two not conflicting forces, but two forces that
exist somewhat independently, but relate to each other. One is the pace of the
progress in AI and ensuring that, you know, the folks working on frontier models are ensuring
those models do benefit humanity. And then there's the sort of geopolitical landscape,
which is, you know, do you want AGI to be created by the freedom, sort of the West,
by democracies, or do you want it to be created by more totalitarian governments? And so I think
the inherent tension for regulators will be a sense of obligation to ensure that, you know,
the technology organizations creating AGI are, in fact, focusing enough on that infity and
humanity, all the other stakeholders who's interests that they're accountable for and ensuring
that the West remains competitive. And I think that's a really nuanced thing. And I think, you know,
my view is it's very important that the West leads in AI, and I'm very proud of the fact
that, you know, Open AI is based here in the United States, and we're investing a lot of the
United States. And I think that's very important. And I also, you know, having sort of seen
the inside of, I think we're really focused on benefiting humanity. So I tend to think that, you know,
it needs to be a multi-stakeholder dialogue, but I think there's a really big risk that some regulations
could have the unintended consequence of slowing down this larger.
conversation. But I don't say that to be dismissive of it either. It's actually just an
impossibly hard problem. And I think you're seeing it play out, as you said, in really
different ways. In Canada, United States, Europe, China, elsewhere. I want to come back to
compute and the dollars involved. So, I mean, on one hand, you have, if I could start an AI
company today by, you know, putting my credit card down using AWS and leveraging their
infrastructure, which they've built, they've spent the hundreds of billions of dollars.
and I get to use it on a time-based model.
On the other hand, you have people like OpenAI
and Microsoft investing tons of money into it
that may be more proprietary.
How do you think about the different models competing?
And then the one that really throws me for a bit of a loop is Facebook.
So Facebook has spent...
You know the matter made a new.
Oh, God.
So meta, I'm like aging myself here.
So meta comes along and, you know, possibly for the good of humanity, but, like, I tend to think Zuck is, like, incredibly smart.
So I don't think he's spending, you know, $100 billion to develop a free model and give it away to society.
How do you think about that in terms of return on capital and return on investment?
It's a really complicated business to be on just given the CAPEX required to build a frontier model.
But let me just start with a couple definitions of terms that I think are useful.
I think most large language models I would call foundation models.
And I like the word foundation because I think it will be foundational to most intelligent systems going forward.
And most people building modern models, particularly if they involve language, image or audio, shouldn't start from building a model from scratch.
They should pick a foundation model, either use it off the shelf.
or fine-tune it.
And so it's truly foundational in many ways.
In the same way, most people don't build their own servers anymore.
They lease them from one of the cloud infrastructure providers.
I think foundation models will be something trained by companies that have a lot of
CAPEX and leased by a broad range of customers who have a broad range of use cases.
And I think that leads in the same way that data center builders having a lot of data
centers enabled you to have the capital scale to build more data centers. I think the same
will largely be true of, you know, building the huge clusters to do training and things like
that. Foundation models, I think, are somewhat distinct from frontier models. And frontier
models, I think it's a term credit to Reed Hoffman, but I may be mistaken on that, but that's
where I heard it from. And these are the models that are usually like the one or two that are
clearly the leading edge, O3, as an example, from Open AI. And these frontier models,
are being built by labs who are trying to build AGI that benefits humanity.
And I think if you're deciding whether you're building a foundation model
and what your business models around it, it's very different business,
then I'm going to go pursue AGI.
Because if you're pursuing AGI, really, there's only one answer,
which is to build and train and move to the next horizon,
because if you can truly build something that is AGI, the economic value is so great,
I think there's a really clear business case there.
If you're pre-training a foundation model that's the fourth best, that's going to cost you
a lot of money.
And the return on that investment is probably fairly questionable because why use your
fourth best large language model versus a frontier model or an open source one for
meta. And as a consequence of that, I think we have probably have too many people building models
right now. There's already been some consolidation, actually, of companies being folded into
Amazon and Microsoft and others. But I do think it will play out a bit like the cloud infrastructure
business where a very small number of companies with very large KAPX budgets are responsible
for both building and operating these data centers. And then developers and in consumers will use
things like chat GPT as a consumer or as a developer, you'll license and rent, you know,
one of these models in the cloud.
How it will play out is a really great question.
You know, I think the, I heard one investor talk about these as like the fastest depreciating
assets of all time.
On the other hand, you know, I, if you look at the revenue scale of something like
an open AI and what I've read about, places like Anthropic, let alone Microsoft and
Amazon, it's pretty incredible as well.
And so you can't really, if you're one of those firms, you can't afford to sit on
the sidelines as the world transforms.
But I would have a hard time personally, like, funding a startup that says I'm going to do
pre-training.
You know, I don't really know, like, what's your differentiation in this marketplace?
And I think a lot of those companies, you're already seeing them consolidate because they
have the cost structure of a pharmaceutical company, but not the business model.
This is just it, though, right?
Like, Open AI has a revenue model around a revenue model.
Microsoft has a revenue model around their AI investments.
They just updated the price of teams with co-pilot.
Amazon has a revenue model around AI, in a sense,
they're getting other people to pay for it through AWS,
and then they're getting the advantages of it at Amazon, too,
from a consumer point of view.
And all the millions of projects, Bezos was doing an interview last week.
that there's every project at Amazon basically has an AI component to it now.
Facebook, on the other hand, has spent all of this money already
and with an endless amount, presumably in sight,
or like not in sight, an endless amount to go,
but they don't have a revenue model specifically around AI,
where it would have been cheaper, obviously, for them to use a different model,
but that would have required presumably giving data away.
I'm just trying to work through it from Zuck's point,
of you if you know i actually will take market as word and you know that post you wrote about
open source i think was very well written and yeah totally people to read it i think that's a
strategy and you know if you look at facebook um you know you've got me saying facebook too so that was
that was what i messed you up yeah you know the company is always really embraced open source
and if i look at really popular things from react to uh you know now the llama models it's
all has been a big part of their strategy to court developers around sort of their ecosystem.
And Mark articulated some of the strategy there. And I'm sure there's elements of commoditizing
your complement. But I also think that, you know, if you can attract developers towards models,
there's a strength. I, you know, I'm not really on the inside there. So I don't really have a
perspective on it other than I actually think it's really great that there's different players with
different incentives all investing so much. And I think it is really furthering the cause of
bringing these amazing tools to society. But a lot changes. I mean, if you look at the
price of GPT40 Mini, you know, it is so much higher quality than like the highest quality
model two years ago and much cheaper. I haven't done the math on it, but it's probably
cheaper to use that than to self-host any of the open source models.
So even the existence of the open source models, it's not free.
I mean, inference costs money.
And so there's a lot of complexity here.
And actually, I have the email even being relatively close to stuff.
Like, I have no idea where things are going.
But, you know, you could talk to a smart engineer and they'll tell you, oh, yeah,
if you built your own servers, you'll spend less than renting them from, say, Amazon
web services.
or Azure. That's sort of true in absolute terms, but this is the fact, like, do you want
someone in your team building servers? Oh, and in fact, if you change the way your service works
and you need a different SKU like you all of a sudden are doing training and you need
Nvidia, you know, H-100s, now all of a sudden you're built servers like this, you know,
asset that's worthless. So I think with a lot of these models, you know, the presence of
open source is incredibly important and I really appreciate it. I also think that the
think like the economics of they are pretty complex because the hardware is very unique. The cost to
serve is much higher. Techniques like distillation have really changed the economics of models whether
or not it's open source or hosted and leased. So it's, I think broadly speaking for developers, it's
kind of an amazing time right now because you have a like a menu of options that's incredibly
wide. And I actually think of it as, you know, just like in cloud computing, you'll end up with
a price performance quality tradeoff. And for any giving engineering talents, they'll have a
different answer. And that's appropriate. And some people use open source Kafka. Some people
work with confluent. Great. You know, like that's just the way these things work, you know.
So you don't think AGI is going to be like a winner take all? You think there's going to be
multiple options that have by definition, whatever the definition is of AGI?
Well, first, I think Open AI, I believe, will play a huge part in it because there's both the technology, which I think Open AI continues to lead on, but also Chat-G-T, which has become synonymous with AI for most consumers.
But more than that, it is the way most people access AI today.
And so one of the interesting things like, what is AGI?
We talked about, you know, opinions on what the definition might be.
But the other question is, like, how do you use that?
Like, what are you, what is the packaging?
And some of intelligence will be simply the outcomes of it, like a discovery of a new drug,
which would be, you know, remarkable and hopefully we can cure some illnesses.
But others will be just how you as an individual access it.
And, you know, most of the people I know, like if they're signing an apartment lease,
we'll put it in a chat, GPD, you get a legal opinion.
And if you get, you know, lab results from your doctor, you can get a second opinion on chat GPT.
Clay and I use the O1 Pro mode for, like, criticizing our strategy at Sierra all the time.
And so for me, what's so remarkable about chat GPT, which was this, you know, quirkily named research preview that has come to be synonymous with AI as I do think that it will be the delivery mechanism for AGI.
when it's produced, and not just because of the many researchers at OpenEI,
but because of the amazing, like, utility it's become from individuals.
And I think that's really neat because, I don't know if it would have been obvious.
If we were having this conversation three years ago, you know,
and you were talking about artificial general intelligence,
I'm not sure either of us would have envisioned something so simple as a form factor
to absorb it that you just talk to it.
So I think it's great, and especially as I think about,
the mission of opening eye, which is to ensure that AGI benefits humanity. What a simple, accessible
form factor. There's free tiers of it. Like, what a kick-ass way to benefit humanity. So I really
think that will be central to what we come as society to define as a AGI. You mentioned using
it at Sierra to critique your business strategy. What do you know about prompting that other people
miss? I mean, you must have the best prompts.
People think that, you know, because I'm affiliated with it.
You're not going, like, here's my strategy.
What do you think?
What are you putting in there?
I often, with the reasoning models, which are slower, will use a faster model first, GBT40, to refine my prompts.
So over the holidays, partly because I was thinking about the future of software engineering,
I've written a lot of compilers in my time.
I'm, like, written enough that I, you know, it's like, a.
It's easy for me.
So I decided to see if I could have 01 Pro mode, generate end-to-end, a compiler, front-end, parsing, the grammar, checking for semantic correctness, generating an intermediate representation, and then using LVM, which is sort of a compiler collection that's very popular to actually do, you know, run, run it all.
And I would spend a lot of time iterating on 4-0 to sort of like refine and make more complete and specific what I was looking for.
And then I would put it into O1 Pro mode, go get my coffee and, you know, come back and get it.
I'm not sure if that's a viable technique, but it's really interesting because I do think in the spirit of AI being the solution to more problems than AI, having a lower latency, simpler model, help refine.
essentially, I like to think
of it as like you're like a product manager
and you're asking an engineer what to do
is your product requirements document
complete and specific enough
and waiting for it
is sometimes slower than and so I like
doing it in stages like that. So that's my
chip. At some point there's probably someone from
open AI listening is going to like roll their eyes but that's
just that's I. Who can I talk to
at Open AI that's like the prompt
ninja? I'm like so curious about this because I've
actually taken recently to
getting
open AI
or chat
TBT I guess
if you want to
call it
I've been getting
chat GBT
to write the
prompt for me
so I'll
prompt it
with I'm prompting
an AI
here are the key things
similar to my technique
I want to accomplish
what would an
excellent prompt
look like
and then I'll copy paste
that prompt
that it gives me
back into the system
but I'm like
I wonder what I'm missing here
right
it's a good technique
I mean there's lots
of engineering techniques
like that, like self-reflection is a technique where you have a model, observe, and critique, you know,
a decision like a chain of thought. So in general, you know, that mechanism of self-reflection
is, I think, a really effective technique. You know, at Sierra, we help companies build customer-facing
AI agents. So if you're setting up a Sona speaker, you'll now chat with an AI. If you're
a serious XM subscriber, you can chat with Harmony, who's their AI to manage your account.
We use all these tricks, you know, self-reflection to detect things like hallucination or decision-making, generating chains of thought for more complex tasks to ensure that it's, you know, you're putting as much compute and cognitive load into important tricks.
So, you know, we're the, there's a whole industry around sort of figuring out how do you exact the like robustness and precision out of these models.
So it's really fun, but changing rapidly.
Hypothetical question, you've been hired to lead or advise a country that wants to become an AI superpower.
What sort of steps would you take?
What sort of policies would you think would help create that?
How would you bring investment from all over the world into that country?
And researchers, right?
So now all of a sudden you're competing.
It's not the United States.
How do you sort of set up a country from first principles all the way back to like, what does
that look like? What are the key variables? Well, I mean, especially, this is definitely
outside of my domain of expertise, but I would say one of the key ingredients to modern AI is
compute, which is a noun that wasn't a noun until recently, but now compute is a noun. And
you know, I do think that's one area where policymakers can, because it involves a lot of things that
touch federal and local governments like power land, and then similarly attracting the capital,
which is immense to finance the real estate, to purchase the compute itself, and then to sort of
operate the data center. And again, there's really immense power requirements for these data
centers as well. And then, you know, it's attracting sort of the right researchers and research
labs to, you know, leverage that, but in general, where there is compute, the research
labs will find you, you know, and I think that's it. And then there's a lot of national
security implications, too, just because, you know, these models are very sensitive. At least
the frontier models are. And so, you know, how you, your place in the geopolitical landscape
is quite important. Like, will research labs and will the U.S. government be comfortable with
training happening there and export restrictions and things like that. But I think a lot of it comes
down to infrastructure as it relates to policy, is my intuition. I think right now so much of
AI is constrained on infrastructure that that is the input to a lot of this stuff. And then there's a
lot around attracting talent and all that. But as I said, you know, you look at the research labs. It's not
that many people, actually. It's a lot, but the compute is a limited resource right now.
That's a really good way to think about it. I think about this for the lens of Canada, right?
Which is like we don't have enough going on in AI. We tend to lose most of our great people to the
states who then go to set up infrastructure here for whatever reason and don't bring it back to Canada.
And I wonder how Canada can compete better. So this is like sort of the lens. I,
like look at these questions through.
How do you see that the next generation of education?
Like if you were setting up a school today from scratch and again, hypothetical, not your
domain of expertise, but like using your lens on AI, how do you think about this?
So like what skills will kids need in the future and what skills do we probably don't need
to teach them anymore that we have been teaching them?
Well, I'll start with the benefits that I think are probably obvious, but I'm incredibly
excited about. I think education can become much more personalized. Oh, totally. Have you seen
synthesis tutor, by the way? No, I have not. Oh, so they developed this synthesis, this AI company,
developed this tutor, which actually teaches kids. And it's so good that El Salvador, the country,
just recently adopted and replaced their teachers. And, like, it'll teach you, but it teaches
you specific to what you're missing. So it's not like every lesson's the same. It's,
it's like, well, you're not understanding this foundational concept.
So it's like K through five or six right now.
That's amazing.
And the results are like off the charts.
Well, it doesn't surprise me.
And I don't actually view it as like necessarily replacing a teacher.
But my view is if you have a teacher with 28 kids in his or her class,
the likelihood that they all learn the same way or learn at the same pace is very unlikely.
And, you know, I can really think of a, say, an English teacher or history teacher,
you're orchestrating their learning journeys through a topic, say AP European history in the
United States, there's a curriculum, they need to learn it, how someone will remember something
or understand the significance of Martin Luther, you know, is very different. And you can, you know,
generate an audio podcast for someone who might be an auditory learner. You can create
cue cards for someone who needs that kind of repetition, repetition.
You can visualize key moments in history for people who just maybe want to more viscerally
appreciate why this was a meaningful event rather than this dry piece of history.
And all of that, as you said, can be personalized to the way you learn and how you learn.
I think it's just incredibly powerful.
And so one of the things I think is neat about AI is it's democratizing access to a lot of
things that used to be fairly exclusive.
A lot of wealthy people, if their child was having trouble in school, would pay for a tutor,
math tutor, science tutor.
And, you know, if you look at kids who are trying to get into, you know, big-name colleges, you know, if you have the means, you'll have someone prep you for the SATs or help you with your college essays, all of that should be democratized if we're doing our jobs well.
And it means that we're not living people's opportunity by their means.
And I think that's just the most American thing ever, Canadian as well.
It's the most incredible thing for humanity.
It's the most incredible thing, humanity.
And so I just think education will change for the positive in so many ways because I actually with my kids walking around when they ask, you know, if you have little kids, they ask why, why, why, and, you know, there's some point a parent just starts making up the answer or being dismissive.
And like, we have chat to PT out.
It's like the best when you're traveling and put on advance.
voice mode and be like, ask away.
100%. And I'm listening too. You know, it's like you're, you live through your children's
curiosity. And my daughter went to high school and came home with Shakespeare for the first
time. And I was, she asked me a question. I was like, I felt this is like total inadequacy.
I was like, I was very bad at this the first time. And then we put it into chat, GPT.
And it was the most thoughtful answer. And she could ask follow up questions. And I actually was,
you know, with her because I was like, oh, I forgot about that. You know, didn't even think about that.
So I just think it's incredible, and I would like to, in public school systems, I think it's really, I think it'll be a really great when public school systems formally adopt these things so that they lean into tools like chat TPT as mechanisms to, like, raise the performance level of their classroom.
And hopefully you'll see it in things like test scores and other things because kids can get the
extra time, even if the school system can't afford it for everyone.
And then most importantly, kids care getting explanations according to their style of learning,
which I think will be quite important as well.
As it relates to skills, it's really hard to predict right now.
And I would say that I do think learning how to learn and learning how to think will continue
to be important.
So I think most of, you know, primary and secondary education shouldn't and is not vocational necessarily.
Some of it is, you know, I took auto shop and all of that, and I'm glad I did.
But I couldn't fix my electric car today with that knowledge.
You know, things change.
And I don't think it needs to be purely, you know, non-vocational.
But, you know, the basics of learning how to think, learning, writing, reading, math, physics.
chemistry, biology, not because you need to memorize it, but understand the mechanisms that
create the world that we live in is quite important. I do think that there's a risk of people
sort of becoming ossified in the tools that they use. So, you know, let's go back to our
discussion of software engineering for a second, but I'll give other examples. You know, if you define your
as a software engineer is how quickly you type into your IDE, the next few years might leave you
behind, you know, because that is no longer a differentiated, you know, part of the software
engineering experience or it will not be. But your judgment as a software engineer will
continue to be incredibly important in your agency and making a decision about what to build,
how to build it, how to architect it, maybe using AI models as a
creative foil. And so I think that just in the same way, if you're an accountant, using Excel
doesn't make you less of an accountant. And just because you didn't, you know, hand craft that
math equation, it doesn't make the results any less valuable to your clients. And so I think we're
going to go through this transformation where I think the tools that we use to create value in
the world will change dramatically. And I think some people who define their
jobs by their ability to use the last generation's tools really, really effectively will be
disrupted. But I think if we can empower people and to re-skill and also broaden the aperture
by which they define the value they're providing to the world, I think a lot of people can make
the transition. The thing that is sort of uncomfortable, not really an education where it's just
earlier in most people's lives. It's just I think the pace of change exceeds that of most
technology transitions. And I think it's unreasonable to expect most people to change the way
to work that quickly. And so I think the next five years, I think will be, you know, for some
jobs really disruptive and tumultuous. But if you take the longer view and you fast forward 25 or
50 years, I'm incredibly optimistic. I think it's the change will require.
from society, from companies, and from individuals,
like an open-mindedness about re-skilling
and reimagining their job to the lens of this
dramatically different new technology.
At what point do we get to...
I mean, we're probably on the cusp of it now
and it's happening in pockets,
but what point do we start solving problems
that humans haven't been able to solve
or eliminating paths that we're on,
maybe with medical research,
that it's like, no, this whole thing,
you've spent $30 billion on, you know, based on this
1972 study that was fabricated, but that one study had all these
derivative studies and like, I'm telling you it's false, you know,
because I can look at it through an objective lens and get rid of these
30 billion.
You're smiling, so.
Oh, no, I just, I hope soon.
I mean, I hope, I mean, I, there was a lot of, there's a, one of the models
I can't remember which one introduced a very long context window,
and there's a lot of people on X over the week.
and putting in their thesis, you know, like grad school thesis in there.
And it was actually critiquing them with, like, surprising levels of fidelity.
And I think we're sort of there, perhaps, with the right tools, but certainly over the next few years.
You know, we talked about what does it mean to generalize AI?
Certainly in the areas of science that are, you know, largely represented through text and digital
technological math being probably the most applicable. There's not really anything keeping AI from
getting really good at math. There's not really an interface to the real world. You don't need to
do a clinical trial to verify something's correct. So I feel a ton of optimism there. It'll be really
interesting in like, you know, areas of like theoretical physics. You'll continue to have the
divide between the applied and the theoretical people. But I think there could be like really interesting
new ideas there and perhaps
some finding logical
inconsistencies with some of the
fashionable theories which has happened many times
over the past few decades
I think we'll get there soon
and actually what's really neat about
is most of the scientists
I know people who are actually doing science
like they're the most excited about these technologies
and they're using them already
and I think that's really neat and I think we're
hopefully going to be
I really hope we see
more breakthroughs in science one of the things i have not an expert in but i've read a lot like a lot
as a amateur about is just the slowdown and scientific breakthroughs over the past you know a few
decades and and some theories that it's because of the degree of specialization that we demand of
grad students and things like that and i hope with you know in general with i um democratizing
access to expertise.
I have a completely personal theory that it will benefit deep generalists in a lot of ways, too,
because your ability to understand a fair amount in a lot of domains and leveraging AI,
knowing where to prompt the AI to go explore and bringing together those domains,
it will start to shift sort of the intellectual power from people who are extremely deep to
people who actually can orchestrate intelligence between lots of different domains or breakthroughs.
I think that'll be really good for society because most scientific breakthroughs aren't.
They tend to be cross-pollinating very important ideas from a lot of different domains, which I think will be really exciting.
How important is the context window?
I think it could be quite important, especially it certainly simplifies working with an AI.
You can just give it everything and instruct it to do something.
And so, and assuming it works, you know, you can extend a context window and it can, the tension can be spread fairly thin and the robustness of the answer can be questionable.
So, but assuming, unless just for argument's sake, you know, perfect robustness, I think it can really simplify the interface to AI.
The not all uses, I also think that we're talking about open source models and APIs.
I also think that most, what I'm excited about in the software industry is not necessarily a large language model with a prompt and a response being the product of AI, but actually end-to-end, close-loose systems that use large language models as pieces of infrastructure.
And I actually think that a lot of the value in software will be that. And for many of those applications, the context window size can matter, but often because you have contextual awareness of the process that you're executing,
context window is a little bit less important.
So I think it matters a lot to intelligence.
You know, there's a, I can't remember someone, one of the, some researcher said, you know,
you put all of human knowledge in the context window and you ask it to invent the next thing.
You know, it's obviously a reductive thought, but interesting.
But I actually, I'm equally excited about sort of the industrial applications of large language models,
sort of like my company, Sierra.
And if you're returning a pair of shoes at a retailer and it's a process that's,
fairly complicated and, you know, is it within the return window? You know, do you want to
return it in store? Do you want to send it? Do you want to print a QR code? Blah, blah, blah, blah.
The orchestration of that is as significant as the models themselves. And I actually think as
we, just like computers, you know, there's going to be a lot of things where computers are a part
of the experience, but it's not like manifesting itself as a computer. So I'm actually equally
excited about those. And I think context window is slightly less important than those applications.
Do you think that the output from AI should be copyrightable or patentable or – let me just take an example.
If I go to the U.S. Patent Office, I download a patent for, let's say, the AeroPress, and I upload it to 01 Pro, and I say – I can't upload it yet because you don't let me do the P.D.S., but I upload it to four.
And so I say, hey, what's the next logical leap that I could patent office?
It would give me back diagrams and an output.
And presumably, if I look at that and I'm like, yeah, that's legit.
I want to file that patent, can I?
I don't know to answer that question.
I'm not an expert in sort of intellectual property, but I think there will be an interesting question of,
was that your idea because you used a tool to do it?
I think the answer is probably yes, that you used the tool to do it.
But I also think that the, in general, like the sort of marginal cost of intelligence will go down a lot.
So a lot of the, you know, I think in general, like, will be in this renaissance of new ideas and intelligence being produced.
And so I think that's broadly a good thing.
And I think, you know, the marginal value of that insight that you had might be lower than it was, you know, and years ago.
What I was hoping you would say is that, you know, that's going to become less and less important because I feel like the patent trolls and all of the stuff that slows down innovation in some ways.
Obviously, like, there's legitimate patents that people infringe on and there should be legal recourse.
But if I could just go and patent like a hundred things a day, because it seems like that should not be allowed.
This is what I'm saying, though.
Well, in general, I think that, you know, companies, you know, I think patents make sense if it's predicting something that's at an active,
use that you invented and you're you're trying to uh you know like the standard you know uh legal
rationale for patents just generating a bunch of ideas and patented it seems destructive to the value
so here's the idea i had last night to counter this because i was like i don't want somebody doing
this uh and i was thinking like what if prior art eliminates patents yeah so i was like what if i
just set off like an instance and just publish it on a website nobody has to read that website here's a
billion ideas. Exactly, but it's like basically patenting like anything,
not patenting, but it's creating prior art for everything. So like you can't
compete on that anymore. I don't know. I was like thinking about that. I thought it was fun.
Reading, playing, learning. Stellist lenses do more than just correct your child's vision.
They slow down the progression of myopia. So your child can continue to discover all
the world has to offer through their own eyes. Light the path to a brighter future,
with Stellist lenses for myopia control.
Learn more at SLR.com
and ask your family eye care professional
for SLR Stellist lenses at your child's next visit.
At Grey Goose, we believe that pleasure is a necessity.
That's why we craft the world's number one
premium vodka in France,
using only three of the finest natural ingredients,
French winter wheat, water from Jean-Sacke and yeast.
With Grey Goose, we
invite you to live in the moment and make time wait.
Seep responsibly.
Tell me about the Google Maps story.
This is like now legend, and I want to hear it from you.
This is my weekend coding.
Is that what you want to hear about?
Yeah.
Yeah, so I'll start with just like the story of Google Maps, the abbreviated version.
We had launched a product at Google called Google Local, which was sort of a yellow pages
search engine, you probably, probably most listeners don't even know what yellow pages are,
but it was a thing back then. And we had licensed maps from MapQuest, which was the dominant
sort of mapping provider at the time. And it was sort of an eyesore on the experience.
And also always felt like it could be a more meaningful part of the kind of local search and
navigation experience on Google. So Larry Page in particular was really pushing us to really
invest more to maps. We found this small.
company with like four people in it, if I'm remembering correctly, started by Lars and Jens
Rasmuson called Where2 Technologies, where they had made a Windows application called Expedition
that was just a beautiful mapping product. It was running on Windows long after it was sort of
out of fashion to make Windows apps, but they were sort of where the technology they're comfortable
with. But they're really, their maps modeled the A to Z maps in the U.K.
and were just beautiful, and they just had a lot of passion for mapping.
So we did a little aqua hire of them and took together the Google Local team
and Lars and Jens's team and said, okay, let's take the good ideas from this Windows app
and the good ideas from Google Local, and like, let's bring them together to make something
completely new.
And that's what became Google Maps.
But there's a couple of idiosyncrasies in the integration.
Because it was a Windows app, it really helped and hurt us.
in a number of ways.
One of the ways it helped us
is the reason why Google Maps
we were able to drag the map
and it was so much more interactive
than any web application that preceded it
was the standard that we needed to hit
from interactivity was set by a native Windows app,
not set by the legacy websites
that we had used at the time.
And I think that by having the goalposts
so far down the field
because they had just started
with this Windows app,
which was sort of a quirk of Lars and Jens, just like technical choices,
we made much bolder technical bets than we would have otherwise.
I think we would have ended up much less interactive had we not started
with that quirky technical sort of decision.
But the other thing was this Windows app, there's a lot of, like,
it's hard to describe the like early 2000s of people who didn't live it,
but like XML was like really in fashion.
So like most things and Windows and other places was like XML and XSLT,
which was a way of transforming XML
into different XML
was the basis of everything.
It was like all of enterprise software
was like XML this, XML that.
So similarly, when we were taking some of these ideas
and putting them in a web browser,
we kind of like went into autopilot
and used like a ton of XML.
And it made everything just like really, really tedious.
And so Google Maps launched
with some really great ideas like the dragable maps.
And we did a bunch of stuff
for the local search technologies.
you could overlay restaurant listings.
It was really great.
It was a really successful launch.
We were like the hot shots within Google afterwards.
But it really started to show its craft.
And we got to this point where we decided we wanted to support the Safari web browser,
which was relatively new at the time.
This was before mobile phones.
And there was much less XML support in Safari than there was an Internet Explorer in Firefox.
And so one of the engineers implemented like a full XSLT transform engine.
in JavaScript to get it to work.
And it was just like shit on top of shit, on top of shit.
And so what was a really elegant fast web application had sort of quickly become something,
you know, there's a lot of dial-up modems at the time and other things.
So like you'd show up to maps and it just was slow and like it just bothered me as like
someone who takes a lot of pride in their craft.
And so I got really energized and like over a weekend and a lot of coffee like rewrote.
it um but it rewrote the whole thing there rewrote yeah more or less the whole thing and it took probably
another week of like you know working through the bugs but yeah i sent it out to the you know the team
after that weekend and it was it was the reason i was able to do it yeah i'm like a decent programmer
but you know you'd also like lived with every bad decision up to that point too so i knew
exactly the output i was going to like i had simulated in my head like if i could do it over again
and this is the way I'd do it.
So by the time I, like, put my hands on the keyboard on, like, you know, Friday night,
it wasn't like I was designing a product.
Like, I knew I had been in every detail of that product since the beginning,
including me the bad decisions, too, the not all the bad decisions.
And so it was just very clear.
I knew what I wanted to accomplish.
And for any, you know, any engineers worked on a big system,
you have the whole system mapped out in your head.
So I knew everything.
And I also knew that, you know,
there's a lot of pride of authorship with engineering and code so I sort of knew I really wanted to finish it over the weekend so that people could use it and see how fast it was and kind of overcome anyone who was like you know you know protective of the code they had written a few months ago and so I really wanted the prototype to go out so I did it and then I didn't it's funny I never talked about it again but I think Paul Bouhite who's was the co-creator of Gmail and and I worked and started friend feed with me
he was on an interview and mentioned this story so now all of a sudden it's like
everyone's talking about it and I was like well thank you Paul
it's a little embarrassed so that people know about it that it was it's a true story and
and and XML is just the worst so did you get a lot of flack from the people who had
built the system you effectively replaced like you were part of that team but everybody
else had so much invested in it even though it was like shit on top of shit on top of
shit you know I wrote a lot of it too so yeah I'm sure there was some around it but
actually, I think good teams want to do great work. And so I think there was a lot of people
constructively dissatisfied with the state of things too. And, you know, I think, you know,
the engineer had written that XSLT transform, I think, was like, you know, a little bit. That's a lot
of work. So you have to throw out a lot of work, which feels bad. But particularly, you know,
Lars and Yens and I, like, we want to make great products. And so I don't think there was a,
you know, at the end of the day, everyone was like, wow, that's.
great. You know, we went from a bundle size of 200k to a bundle size of 20K, and it was a lot
faster and better. So, you know, broadly speaking, I think good engineering cultures, you don't want
a culture of, you know, ready, fire aim, but I also think you just need to be really outcomes-oriented.
And I think people, if they become, they'd start to treat their code as too precious. It can
really impede forward progress. And, you know, I'll just take, like, my understanding,
is like a lot of the early self-driving car software was a lot of hand-coded heuristics and
rules and you know a lot of smart people think that eventually will probably be a more monolithic
model that encodes many of the same rules you have to throw out a lot of code in that transition
but it doesn't mean it's not the right thing to do and so I think in general yeah there might
have been some feathers ruffled but at the end of the day everyone's like that's faster and better like
let's let's do it you know which is I think the right decision that's awesome and to give you
another hypothetical. I want you to share your inner monologue with me as you think through it.
So if I told you you have to put 100% of your net worth into a public company today,
and you couldn't touch it for at least 20 years, what company would you invest in? And like,
walk me through your thinking. I literally don't know how to answer that question.
How would you think about it without giving me an answer? Like, what are the...
Yeah, it's a good question. First of all, I'll give you how I'd think about it, but I'm so...
Having not been a public company CEO for a couple years, I blissfully don't pay attention as much to the public markets.
And in particular right now, it's obviously valuations have gone up a lot.
So there's a – but because it's a long-term question, maybe that doesn't matter.
I think what I'd be thinking about right now is over the next 20 years, like what are the parts of the economy that will most benefit from this current wave of AI?
That's not the only way to invest over 20-year period,
but certainly it's a domain that I understand.
And in particular, you know, I mentioned that talk
I heard a snippet of from Tyler Cohen,
which is like it will probably, AI will probably benefit
different parts of the economy disproportionately.
There will be some parts of the economy
that can essentially where intelligence is a limiting factor
to its growth and where you can absorb
almost arbitrary levels of intelligence
and generate almost arbitrary levels of growth.
Obviously, there's limits to all of this
just because you change one part of the economy.
It impacts other parts of the economy,
and that was what Tyler's point was in his talk.
But I would probably think about that
because I think that over a 20-year period,
there are certain parts of society
that won't be able to change extremely rapidly,
but there will be some parts that probably will,
and it'll probably be domains where intelligence
is the scarce resource right now.
And then I would probably try to find companies
that will disproportionately benefit from it.
And I assume this is why, like,
NVIDIA's stock is so high right now
because if you want to sort of get downstream,
you know, NVIDIA will probably benefit
from all of the investments in AI.
I'm not sure I would do that over a 20-year period,
just assuming that the infrastructure will shift.
So I don't have an intelligent answer,
but that's the way I would think about it
if we were doing that exercise.
I love that.
Where do you think?
like what's your intuition say about what areas of the economy are limited by intelligence
and not just economy, I mean, perhaps politicians might be limited by this and aid and benefit
from, in which case countries could benefit enormously from AI and unlock growth and
potential in their economy. But I think maybe just to scope the question, like what areas of
the economy do you think are limited by intelligence or workers, like smart workers, in which
case like that's another limit of intelligence yeah i mean uh two that are i think probably going to
benefit a lot are technology and finance um you know where you're you know if you can make
better financial decisions than competitors you'll generate outsized returns and that's why over
the past you know 30 years you know of machine learning um you know uh hedge funds and financial
services institutions, everything from fraud prevention to true investment strategies,
it's already been an area of domain, domain of investment. Software, similar as we talked
about, I think that at some point, we will be, we will no longer be supply constrained in software,
but we're not anywhere close to it right now. And you're taking something that has always been
the scarce resource, which is software engineers, and you're making it not scarce. And I think as a
consequence. You just think of like how much can that industry grow? We don't know. But we've been
so constrained on software engineering as a resource. Who knows over the next 20 years, but we'll
find out where the limits are. But to me, intellectually, there's just a ton of growth there.
And then broadly, I think areas of like processing information are areas that will really benefit
quite a bit here. And so that, and I think the thing that I would think about over 20
period is like second and third order effects, which is why I don't have an intelligent
answer. And if you're asking me to put all my money in something, I would think about it for a
while, probably use O1 Pro a little bit to help me. But, you know, because you can end up generated
a bunch of growth in the short term, but then, you know, if everyone does it, it commoditizes the
whole industry, you know, type of thing. So, you know, there used to be, you know, before the
introduction of the freezer, ice was like a really expensive thing and now it's free, you know.
And so I think it is really important to actually think through those if you're talking
in a time frame of like 20 years.
And that's why having not thought about this question ahead of time, you could be quite
simplistic elsewhere.
But I would say software and finance are areas that I think stand to reason should benefit
quite a bit.
I love that response.
How do you balance having a young family with also running a startup again?
I work a lot.
I don't.
I really care.
and love care about and love working.
So one thing is that I, well, there's always tradeoffs in life.
If I didn't love working, I wouldn't do it as much as I do.
But I just love to create things and love to have an impact.
And so I like jump out of bed in the morning and work out, go to work.
And then it's been telling my family broadly, probably being honest.
first I'm not perfect
but second I don't have a ton of hobbies
you know I basically work and spend time with my family
the first time we talked to you saw the couple
guitars in my background
I haven't picked one of those up in a while
I mean I literally pick it up occasionally
but I you know do not devote any time into it
and I don't regret that either like I am so
passionate about what we're building at Sira
I'm so passionate about opening I'm so
love my family so much I don't really have any regrets
about it but I basically just
life is all about where do you spend your time and mine is at work and with family and so that's how
I do it I don't know if I'm particularly balanced but I don't strive to be either I really take a lot
of pride and I love I love to work having sold the companies you started twice how does that influence
what you think of Sierra like are you thinking like oh I'm building this in order to sell it or
do you think differently like this is my life's work I'm building this with that's not going to
happen? I absolutely
intend Sierra to be
an enduring company and an independent
company. But to be honest,
every entrepreneur with every company starts
that way. And so
I'm really grateful
for both Facebook and Salesforce for having
acquired my previous companies and hopefully
I had an impact about those companies. But
you don't start off, well, at least I never
started off saying, hey, I want to make
a company to sell it.
But I actually
actually think with Sierra we have just a ton of traction in the marketplace. I really do think
Sierra is a leader in helping consumer brands build customer facing AI agents. And I'm really
proud of that. So I really see a path to that. And I joke with Clay, I want to be an old man
sitting on his porch, you know, complaining how the next generation of leaders at Sierra don't
listen to us anymore. You know, I want this to be something that not only is enduring, but outlives
me. And I think just actually, I don't think we've ever talked about this, but it was,
really interesting moment for me when
Google went from its one building in Mountain View to its first
corporate campus. We moved into the Silicon
Graphics campus, which was right over near Shoreline Boulevard
in Mountain View. And
SGI had been a really successful company enough to build a
campus. And when we, it was actually quite awkward, we moved
into like half the campus. They were still in half and they're like,
we're this up-and-coming company, they're declining. And then
when Facebook, when we moved,
out of the second building.
We were in Palo Alto,
a slightly larger building.
I think we leased it from HP.
But when we finally got a campus,
it was from Sun Microsystems
who had gone through an Oracle acquisition
and had been sort of on the decline.
And it was interesting to me
because both SGI and Sun
had been started and grown
to prominence in my lifetime.
Obviously, it was maybe like
a little younger, obviously,
but in my lifetime,
enough to build a whole corporate campus
and then declined
fast enough to sell that corporate campus to a new software company.
And for me, it was just so interesting to have done that twice to move into like a, you know, a used campus for the previous owners.
It was a very stark reminder that technology companies aren't entitled to their future success.
And I think we'll see this actually now with AI.
AI, I think, will change the landscape of software to be tools of productivity.
to agents that actually accomplish tasks.
And I think it will help some companies for whom that's a amplifies their existing value
proposition and it will really hurt others where it will essentially the seat-based kind
of model of legacy software will wane very quickly and then really harm them.
And so when I think about what it means to build a company that's enduring, that is a really,
really tall task in my mind right now because it means not only making something that's
financially enduring over the next 10 years, but setting up a culture where a company can actually
evolve to meet the changing demands of society and technology when it's changing at a pace
that is like unprecedented history. So I think it's one of the most fun business challenges
of all time. And I think it has as much to do with culture as it has to do with technology because
every line of code in Sierra today will be completely different, you know, probably five years
from now, let alone 30 years from now. And I think that's really exciting. So when I think about it,
I just get so much energy because it's incredibly hard and it's harder now that it's ever been
to do something that lasts beyond you. But that I think is the ultimate measure of a company.
You mentioned AI agents. How would you define that? What's an agent? I'll define a more
broadly, and then I'll tell you how we think about it, Sierra, which is a more narrow view of it.
The word agent comes from agency, and I think it means affording a software, the opportunity
to reason and make decisions autonomously.
And I think that's really all it means to me, and I think there's lots of different
applications of it.
The three categories that I think are meaningful, and I'll end with the Sierra one, just so I
can talk about it a little more, but one is personal agents. So I do think that most people will
have probably one, but maybe a couple AI agents that they use on a daily basis that are
essentially amplifying themselves as an individual. You can do the rote things like help you
triage your email to help you schedule a vacation. You know, you're flying back to Edmonton
and help you arrange your travel.
Two more complex things.
Like, you know, I'm going to go ask my boss for promotion, like help me role play.
And, you know, I'm setting up my resume for this job, help me do that to, I'm applying
for a new job, help me find companies I haven't thought of that I should be applying to.
And I think these agents will be really powerful.
I think it might be a really hard product to build because when you think about all the different
services and people you interact with every day, it's kind of.
of everything. So it's not, it has to generalize a lot to be useful to you. And because of the
personal privacy and things like that, it has to work really well for you to trust it. So I think
it's going to take a while to go, I think it'll be a lot of demos. I think it'll take a while
to be robust. The second category of agent is, I would say, really filling in a persona
within a company. So a coding agent, a paralegal agent, a analyst agent.
I think these already exist.
I mentioned Cursor.
There's a company called Harvey that makes a legal agent.
I'm sure there's a bunch in the analyst space.
These do a job, and they're more narrow, but they're really commercially valuable because
most companies hire people or consultants that do those things already, like analyze the contracts
of your supply chain, right?
That's kind of a rote kind of law, but it's really important, and AI can do it really well.
So I think that's why this is the area of the economy that I think is really exciting.
And I'm really excited about all the startups in this space because you're essentially taking what used to be a combination of people and software and really making something that solves a problem.
And by narrowing the domain of autonomy, you can have more robust guardrails and even with current models actually achieve something that's effective enough to be commercially viable today.
and by the way it changes the total addressable market of these models too like
I don't know what the total addressable market of legal software was three years ago
but it couldn't have been that big I couldn't tell you like a legal software company
I probably should I just can't think of one but if you think about the money we spend on
lawyers that's a lot and so you end up where you're broadening the addressable market
quite a lot the domain we're in I think is somewhat special which is
a company's branded customer-facing agent.
And the reason why I think it's, one could argue,
we're sort of helping with customer service,
which is a persona, a role.
But I do think it's broader than that
because if you think about a website,
you know, like your insurance company's website,
try to list all the things you can do on it.
You can look up the stock quote.
You can look up the management team.
You can compare their insurance company
to all their competitors.
You can file a claim.
You can, you know, buy, you can bundle your home and auto.
You can add a member of your family to your premium.
There's a million things you can do on it.
Essentially, over the past 30 years, websites, a company's website, singular,
has come to be the universe of everything that you can do with that company.
I like to think it was like the digital instantiation of the company.
And that's what we're helping our customers do at Sirius,
help them build a conversational AI that does all of that.
So, you know, most of our customers start with customer service,
and it's a great application because no one likes to wait on hold
and having something that has perfect access to information
is multilingual and empathetic is just amazing.
But, you know, when you put a conversational AI
as your digital front door, people will say anything they want to it.
And we're now doing product discovery, consider purchases, going back to insurance example.
Hey, you know, I've got a 15-year-old daughter.
I really am concerned about the cost of her premium until she grows up.
Tell me which plan I should be on.
Tell me why you'll be better than your competitors.
That's a really complex interaction, right?
That's not something that can you make a web page that does that?
No, but that's a great conversation.
And so we really aspire that when you encounter.
a branded agent in the wild, we want Sierra to be the platform that powers it.
And it's super important because there was a case, at least in Canada, where an AI agent for Air Canada
hallucinated a bereavement policy. Right, but they were found liable to hold themselves to what
the agent said. Yeah, I mean, it turns out... And it was an AI agent. There was no human involved
in the whole thing. Well, look, it's one thing if chat GPT hallucinate something about your brand.
It's another if your AI agent, who they say something about your brand.
So the bar just gets higher.
So the robustness of these agents, the guardrails, everything is more important when it's
yours and it has your brand on it.
And so it's harder, but I also, I'm just so excited for it because this is a little
overly intellectual, but I really like the framing.
If you think about a modern website or mobile app, it's essentially you've created a
directory of functionality from which you can choose. But the main person with agency and that is
the creator of the website. Like what are the universe of options that you can do? When you have
an AI agent represent your brand, the agency goes to the customer. They can express their problem
any way they want in a multifaceted way. And so it means that like your customer experience goes
from the enumerated set of functionality you've decided upon your website to whatever your
customers ask. And then, you know, you can decide how to fulfill those requests or whether
you want to. Yeah, yeah. But I think it would really change the dynamic to be really empowering
to consumers. As you said, I mean, the reason that Air Canada case is the reason we exist.
You know, companies, if they try to build this themselves, there is a lot of ways you can
shoot yourself in the foot. But in particular, too, your customer experience should not be wedded
to one model, let alone even this current generation of models.
So with Sierra, you can define your customer experience once in a way that's abstracted
from all of the technology.
And it can be a chat.
It can be, you can call you on the phone.
It can be all of those things.
And as new models and new technology comes out, our platform just gets better.
But you're not like re-implementing your customer experience.
And I think that's really important because we were talking about what's happened over the past
two years. Can you imagine if you're a consumer brand like ADT home security and thinking about like how can
you maintain your AI agent in the face of all of that, right? It's just not even, it's not tenable. I mean,
it's not what you do as ADT. So they've worked with us to build their AI agent. How do you fend off
complacency? Like a lot of these companies and maybe not in tech specifically, but they get big,
they get dominant, and then they take their foot off the gas. And that open.
opens the door to competitors.
And there's like a natural entropy almost to bureaucracy in some of these companies that
and the bureaucracy serves the seeds of failure and competition.
How do you, how do you fend that off constantly?
It is a really challenging thing to do at a company.
One of the, there's two things that I've observed that I think manifest as corporate complacency.
One is bureaucracy.
And I think the root of bureaucracy is often, when something goes wrong, companies introduce a process to fix it.
And over those like sequence of 30 years, the layered sum of all of those processes that were all created for good reason with good intentions, end up being a bureaucratic sort of machine where the reason.
the reasons for many of the rules and processes
are rarely even remembered by the organization
but it creates this sort of
natural inertia
sometimes that inertia can be good
you know it's like you know if you end up with
there's definitely been stories of executives coming in
and ready fire aim new strategies that backfire
massively but often it can mean
in the face of a technology shift or a new competitor
you just can't move fast enough to address it
the second thing that I think is more subtle is as a company grows in size,
often its internal narrative can be stronger than the truth from customers.
I remember one time when this sort of peak of the smartphone wars and I end up visiting
a friend on Microsoft's campus.
And I got off the plane in Seattle-Tacoma Airport, drove into Redmond, went onto the campus.
and all of a sudden everyone I saw was using Windows phones.
I assume it must have been a requirement or formal or social.
You were definitely uncool if you were using anything else.
And from my perspective, at the time, like the war had already been lost.
Yeah.
Like it was definitely a two-horse race between Apple and Google on iOS and Android.
And I remember sitting in the lobby waiting for my first.
friend to get me from the security check-in.
And I made a comment, like, it wasn't a confrontation about, but I made a comment to
someone who was at Microsoft, I was like, you know, something along the lines, are you
required to use Windows phones, how these others? And I just sort of like curious. And then I got a
really bold answer, which is like, yeah, we're going to win. Like, we're taking over the
smartphone market. And I was like, I didn't say anything because it was like a little socially
awkward of like, no, you're not. Like, you lost like four years ago.
But there's a process, there's something that's happening that's preventing you from getting reality.
Well, and that's the thing is if you think about it, if anyone, if you've ever worked for like a large company, you know, you're, when you work at a small company, you care about your customers and your competitors and you feel every bump in the road, when you're a, you know, junior vice president of whatever and you're, you know, eight levels below your, you know, CEO.
and you have a set of objectives and results.
You might be focused as I want to go from junior vice president to senior vice president.
That's what success looks like for me.
And you end up with this sort of myopic focus on this internal world.
In the same way your kids will focus on, you know, the social dynamics of their high school,
not the world outside of it.
And it's probably rational, by the way, because like, you know,
probably their social life is more determined by those.
you know 1,000 kids in their high school and it's like all the things outside but this is that's the
life of a person inside of these big places and so you end up where uh you know if you have a very
senior head of product who's like are this competitor says they're faster but this next version
we're so much better and then everyone says and all of a sudden that's like the windows phone is
going to win that's what everyone says and and you truly believe it because everyone you meet
says the same thing and you end up reflecting, you know, customer anecdotes through that lens
and you end up with this sort of reality distortion field manifested from the sum of this
sort of myopic storytelling that exists within companies. What's interesting about that is like,
you know, the ability for a culture to believe in something is actually a great strength of a culture,
but it can lead to this as well. And so the combination of bureaucracy and inaccurate stories,
storytelling, I think, is the reason why companies sort of die. And it's really remarkable to look at, you know, the BlackBerrys of the world or the TiVos or that, you know, you can really, you know, as the plane is crashing, like, tell the story that you're not. And, and, and then similar, as I said, like, culturally, you can still have, like, the person in the back of that crashing plane being like, when am I going to get promoted to SVP?
and you're like, what the, you know.
Yeah.
And that's, I mean, this is like, I mean, I've seen it a hundred times.
And so I think it really comes down to leadership, you know, and I think that one of the things that most great companies have is they are obsessed with their customers.
And I think the free market doesn't lie.
And so I think the one of the most important things I think for any like enduring culture, particularly in an industry that changes as rapidly as software is how close are your employees to customers and how much.
can customer, like the direct voice of your customers, be a part of your decision-making?
And that is something that I think you need to constantly work out
because that, you know, person, employee number 30,000, 462, you know,
how does he or she actually directly hear from customers?
It's not actually a simple question to answer.
Is it direct? Is it filtered? How many filters are there?
That's exactly right. And then I think the,
other part on leadership is, you know, we talk about bureaucracy is process is there to
serve the needs of the business. And often mid-level managers don't get credit for removing
process. They often are held accountable for things going wrong. And I think it really takes
top-down leadership to, you know, remove bureaucracy. And it is not always comfortable, you know,
when companies remove spans of control or all the people impacted, well, it's like antibodies.
And for good reason, it makes sense.
Their lives are negatively impacted or whatever it is.
But it almost has to come from the top because you need to give air cover.
Almost certainly something will go wrong, by the way.
I mean, like processes usually exist for a reason.
But when they accumulate without end, you end up with bureaucracy.
So those are the two things that I always, and you could smell it when you go into a really bureaucratic company, the inaccurate storytelling, the process over outcomes.
And it's just, it sort of sucks the energy out of you and you feel it.
That's a great answer.
We always end these interviews with the exact same question, which is, what is success for you?
Success for me.
We talked about how I spend my time with my family at work is, you know, having a happy, healthy family.
and being able to work with my co-founder, Clay, for the rest of my life,
making Sierra into an enduring company.
That would be success for me.
Thanks for listening and learning with us.
The Farnham Street blog is where you can learn more about my new book,
Clear Thinking, Turning Ordinary Moments into Extraordinary Results.
It's a transformative guide that hands you the tools to master your fate,
sharpen your decision-making and set yourself up for unparalleled success.
Learn more at fs.blog slash clear.
Until next time.