The Knowledge Project with Shane Parrish - Bret Taylor: A Vision for AI’s Next Frontier

Episode Date: April 15, 2025

What happens when one of the most legendary minds in tech delves deep into the real workings of modern AI? A 2-hour long masterclass that you don’t want to miss.   Bret Taylor, current chairman of ...OpenAI, unpacks why AI is transforming software engineering forever, how founders can survive acquisition (he’s done it twice), and why the true bottlenecks in AI aren’t what most think. Drawing on his extensive experiences at Facebook, Google, Twitter and more, he explains why the next phase of AI won’t just be about building better models—it’s about creating entirely new ways for us to work with them. Bret exposes the reality gap between what AI insiders understand and what everyone else believes. Listen now to recalibrate your thinking before your competitors do.  (00:02:46) Aha Moments with AI (00:04:43) Founders Working for Founders (00:07:59) Acquisition Process (00:14:14) The Role of a Board (00:17:05) Founder Mode (00:20:29) Engineers as Leaders (00:24:54) Applying First Principles in Business (00:28:43) The Future of Software Engineering (00:35:11) Efficiency and Verification of AI-Generated Code (00:36:46) The Future of Software Development (00:37:24) Defining AGI (00:47:03) AI Self-Improvement? (00:47:58) Safety Measures and Supervision in AI (00:49:47) Benefiting Humanity and AI Safety (00:54:06) Regulation and Geopolitical Landscape in AI (00:55:58) Foundation Models and Frontier Models (01:01:06) Economics and Open Source Models (01:05:18) AI and AGI Accessibility (01:07:42) Optimizing AI Prompts (01:11:18) Creating an AI Superpower (01:14:12) Future of Education and AI (01:19:34) The Impact of AI on Job Roles (01:21:58) AI in Problem-Solving and Research (01:25:24) Importance of AI Context Window (01:27:37) AI Output and Intellectual Property (01:30:09) Google Maps Launch and Challenges (01:37:57) Long-Term Investment in AI (01:43:02) Balancing Work and Family Life (01:44:25) Building Sierra as an Enduring Company (01:45:38) Lessons from Tech Company Lifecycles (01:48:31) Definition and Applications of AI Agents (01:53:56) Challenges and Importance of Branded AI Agents (01:56:28) Fending Off Complacency in Companies (02:01:21) Customer Obsession and Leadership in Companies Bret Taylor is currently the Chairman of OpenAI and CEO of Sierra. Previously, he was the CTO of Facebook, Chairman of the board for X, and the Co-CEO of Salesforce.  Newsletter - The Brain Food newsletter delivers actionable insights and thoughtful ideas every Sunday. It takes 5 minutes to read, and it’s completely free. Learn more and sign up at fs.blog/newsletter Upgrade — If you want to hear my thoughts and reflections at the end of the episode, join our membership: ⁠⁠⁠⁠⁠⁠⁠fs.blog/membership⁠⁠ and get your own private feed. Watch on YouTube: @tkppodcast Learn more about your ad choices. Visit megaphone.fm/adchoices

Transcript
Discussion (0)
Starting point is 00:00:00 Technology companies aren't entitled to their future success. AI, I think, will change the landscape of software, and I think it will help some companies, and it will really hurt others. And so when I think about what it means to build a company that's enduring, that is a really, really tall task in my mind right now, because it means not only making something that's financially enduring over the next 10 years, but setting up a culture where a company can actually evolve to meet the changing demands of society and technology when it's changing at a pace that is like
Starting point is 00:00:35 unprecedented history. So I think it's one of the most fun business challenges of all time. I just get so much energy because it's incredibly hard and it's harder now that it's ever been to do something that lasts beyond you. But that I think is the ultimate measure of a company. podcast. I'm your host, Shane Parrish. In a world where knowledge is power, this podcast is your toolkit for mastering the best of what other people have already figured out. If you want to take your learning the next level, consider joining our membership program at fs.blog slash membership. As a member, you'll get my personal reflections at the end of every episode, early access to episodes, no ads, including this, exclusive content, hand-edited
Starting point is 00:01:28 transcripts and so much more. Check out the link in the show notes for more. Six months after Brett Taylor realized AI was about to change everything, he walked away from his co-CEO job at Salesforce to start from scratch. That's how massive this shift really is. The mastermind behind Google Maps and the former chief technology officer at Facebook, Brett reveals the brutal truce about leadership, AI and what it really takes to build something that endures long after you've reached the top. Brett's led some of the most influential companies in tech and seen exactly what makes businesses scale, what kills them from within, and why most founders don't survive their own success. In this conversation, you'll discover why so many companies are already on life
Starting point is 00:02:12 support without realizing it, how first principles thinking separates the next wave of winners from everyone else and the hidden reason most acquisitions fail. We'll explore why AI is bigger than anyone suspects, plus the mindset shift that turns great engineers into exceptional CEOs. Whether you're a founder, an operator, or simply someone who wants to think sharper, this episode will change how you see your business, technology, and the future. It's time to listen and learn. What was your first real aha moment with AI where you realized, holy shit, this is going to be huge? I had two separate aha moments, one that I don't think I really appreciated how huge it would be,
Starting point is 00:03:00 but it kind of reset my expectation, which was the launch of Dolly in the summer of 22, is that right? That might be off a year, but I think summer of 22, and the avocado chair that they generated. And I had been, well, my background is in computer science and pretty technically deep. I hadn't been paying attention to large language models. I just didn't follow the progress after the Transformers paper. And I saw that and my reaction was I had no idea computers could do that. And that particular launch, you know, seeing a generated image of an avocado chair, I don't think I extrapolated to what, you know, where we are now.
Starting point is 00:03:46 But it, for me, shook me and realized I'd need to pay more. attention to this space and Open AI specifically than I had been. I think I had, that was the moment where I realized, like, I clearly have been not paying attention to something significant. And then it was, you know, six months later, coincidentally, like the month after I left Salesforce, the chat CBD came out and before it became a phenomenon, though it did so quickly, but I was already, you know, plugged into it. And I was, from that on, you know, I could not stop thinking about it.
Starting point is 00:04:21 But that avocado chair, I don't know why. I think it was the, there was a bit of an emotional moment where you saw a computer doing something that wasn't just rule-based, but creative. And the idea of a computer doing something, creating something from scratch was, well, it doesn't seem so novel. You know, a few years later, just blew my mind at the time. One of the unique things about you is that you've started companies, you've been acquired by Facebook and Salesforce. Inside those companies, you rose up to be the CTO at Facebook, the co-CEO at Salesforce. Talk to me about founders working for founders and founders working within a company. Yeah, it's a very challenging transition for a lot of founders to make.
Starting point is 00:05:10 I think there's lots of examples of acquisitions that have been really transformative from a business standpoint. I think YouTube, Instagram, being two of the more prominent that have clearly changed the shape of the acquiring company. But even in those cases, you know, the founders didn't stay around that long. And that's maybe a little unfair. Stick around for a little bit. I think the interesting thing about being a founder is it's not just building a business, but it's very much your identity. And I think it's very hard for people who aren't founders to experience it. You take everything very personally, you know, from the product to the customers, to the press, to your competitors, the both inner and outer measures of success.
Starting point is 00:05:57 And I think when you go to being acquired, there's a business aspect to it. And can you operate within a larger company? But that's intertwined with a sense of identity. You go from being the founder of a company and the CEO of a company or CTO of a company, whatever your title happens to be, as one of the co-founders. to be in a part of a larger organization and to fully embrace that, you actually need to change your identity. You need to go from being, you know, the head of Instagram, in my case, the head of QIP to being an employee of Salesforce or going from being the CEO of Friend Feed to being an employee of Facebook. And what I've observed is it's that identity shift is a prerequisite for most
Starting point is 00:06:39 of the other things. It's not simply your ability to handle the politics and bureaucracy of a bigger company or to navigate a new structure. I actually think most founders don't make that leap where they actually identify with that new thing. It's even harder for some of the employees too because most of the time in an acquisition, an employee of an acquired company didn't choose that path. And in fact, they chose to work for a different company and they, you know, the acquisition determined a different outcome. And that's why integrating acquisitions is so nuanced. And I would say that having the experience of having been acquired, you know, before and having acquired some companies before when I got to Salesforce, I really tried to be self-aware about that and really
Starting point is 00:07:23 tried to, you know, be a part of Salesforce, you know, and tried to shift my identity and not be a single-issue voter around QIP. You know, I'd really try to embrace it. And I think it's really hard for some founders do. And some founders don't want to, honestly, you know, they maybe cash the check and and you know that's the it's more of a transactional relationship I I really actually am so grateful for the experience of having been at Facebook and Salesforce I learned so much so but it really took a lot of effort on my part to just um transform my perception of myself and who I am to get that value out of the company that acquired us how did you how did it change how you did acquisitions at Salesforce you guys did a lot of acquisitions what you were there
Starting point is 00:08:06 and you're acquiring founders and sort of startups, and I think Slack was while you were there too. How did that change how you went about integrating that company into the Salesforce culture? I'll talk abstract about talking about some specific acquisitions too, but first I think I tried to approach it with more empathy and more realism. You know, one of the nuanced parts about acquisitions is
Starting point is 00:08:31 there's the period of doing the acquisition, there's sort of the period after you've decided to do it of doing due diligence, and then there's a period when it's done, and you're integrating the company and sort of the period after. One of the things that I have observed is that companies doing acquisitions, often the part of deciding to do it is a bit of a mutual sales process. You're trying to find a fair value for the company, and there's some back and forth there.
Starting point is 00:09:02 But at the end of the day, there's usually some objective measure of that influenced by a lot of factors, but there's some fair value of that. But what you're trying to do is what are, and corporate speak, would be synergies, but like, why do this? Why is one plus one greater than two? You know, that's why you do an acquisition just from first principles. It's often an exercise in storytelling. You know, you bring this product together with our product, and customers will, you know, find the whole greater than the sum of its parts. This team applies, to our sales channel or if you're a Google acquisition, you know, imagine the traffic we can drive to this product experience, you know, in the case of something like an Instagram,
Starting point is 00:09:45 imagine our ad sales team attached to your amazing product and how quickly we can help you realize that value, whatever it might be. I find that people, because there's sort of a craft of storytelling, for both sides to come to the same conclusion that they should do this acquisition, sometimes either simplifies or sugarcoats like some of the realities of it, you know, little things like, you know, how much control will the founding team of the acquired company have over those decisions? Will it be operated as a standalone business unit or will your team be sort of broken up into functional groups within the larger company? and it's sort of those little, they're not little, but those I'll say boring but important things
Starting point is 00:10:34 that often people don't talk enough about. And you don't need to figure out every part of an acquisition to make it successful, but often you can end up running into like true third rails that you didn't find because you were having the storytelling discussions rather than getting down to brass tax about how things will work and what's important. The other thing that I think is really important is being really clear what success looks like. And, you know, I I think sometimes it's a business outcome, sometimes it's a product goal, but I found that if you went to most of the larger acquisitions in the Valley, and you, two weeks after it was closed, interviewed the management team of the acquiring company and the acquired company,
Starting point is 00:11:18 and you asked them, like, what does success look like two years from now? My guess is, like, 80% of the time you get different answers. And I think it goes back to this sort of storytelling thing where you're talking about the benefits of the acquisition about talking about like, what does success look like? So I really tried to approach it. I tried to pull forward some harder conversations when I'm doing acquisitions or even when I'm being acquired since it's happened to be not twice. So that, you know, when you're approaching it, you'd not only get the, hey, why is one plus one equal gradient to two? Everything's going to be awesome, you know. but no for real like what you know what does success look like here and then you know as a founder
Starting point is 00:11:58 your job of an acquired company is to tell your team that and align your team to that and I think founders don't take on enough accountability towards making these acquisitions successful as I think they should and and it goes back to again a certain uh naivete you know it's like you're you're not your company anymore you're part of something larger and I think you know successful ones work when everyone embraces, embraces that. What point in the acquisition process is that conversation? Is that after we've signed our binding, you know, sort of commitment? Or is it we should have that conversation before so I know what I'm walking into?
Starting point is 00:12:36 My personal take is it's not something you have, you have to get to the point where the two parties want to merge, you know. And that's obviously a financial decision, particularly if it's like a public company. there's a board and shareholders. Most acquisitions in the Valley are a larger firm acquiring a private firm. That's not all of them, but I would say that's the vast majority. And in those cases, there's often a qualitative threshold where someone's like, yeah, let's do this. We kind of have the high-level terms, sometimes a term sheet, you know, formally. I think it's right after that.
Starting point is 00:13:09 So where people have really committed to the key things, how much value, why are we doing this, the big stuff. and there's usually, you know, many, lots of lawyers being paid lots of money to turn those term sheets into, you know, more complete set of documents, usually more complete due diligence, stuff like that. That's a, there's an awkward waiting period there. And that's a time, I think, where, like, the strategic decision makers in those moments can get together and say, let's talk through what this really means. And the nice part about having them for all parties is you've kind of made the commitment to
Starting point is 00:13:46 each other so it's you've I think you have more social permission to have real conversations at that point but you also haven't consummated the relationship you know and so there's a the power imbalance isn't totally there and and you can really talk through it and it also I think engender's trust just because by having a harder conversations in those moments you're learning how to have real conversations and learning how each other works so that's my my personal opinion when to have it So you mentioned the board, you've been on the board of Shopify, you're on the board of OpenAI, you're a founder. What's the role of a board and how is it different when you're on the board of a founder-led company? With Amex Platinum, $400 in annual credits for travel and dining means you not only satisfy your travel bug, but your taste buds too.
Starting point is 00:14:36 That's the powerful backing of Amex. Conditions apply. Hit pause on whatever you're listening. to and hit play on your next adventure. This fall get double points on every qualified stay. Life's the trip. Make the most of it at BestWestern. Visit bestwestern.com for complete terms and conditions. I really like being involved in a board, and I've been involved in multiple boards
Starting point is 00:15:04 because I think I am an operator through and through. I probably self-identifies an engineer first more than anything else, and I love to build. learning how to be an advisor is a very different vantage point that I think you see how other companies operate and you also learn how to have an impact and add value without doing it yourself and it's a very and I've really I think become a better leader you know having learned to do that I have really only joined boards that were led by founders because typically I think they, you can speak to them, but I think they sought me out because I'm a founder and I like working with founder-led companies. I think the founders, I'm sure there's lots of studies on this,
Starting point is 00:15:54 but I think founders drive better outcomes for companies. There's a, I think founders tend to have permission to make bolder, more disruptive decisions about their business than a professional manager. There's exceptions like Sotcha, I think, is, you know, one of the greatest, if not the greatest CEO of, you know, our generation and as a professional manager. But, you know, you look at everyone from Toby Lukie to Mark Benioff to Mark Zuckerberg to Sam at OpenAI. And I think when you have founded a company, all your stakeholders, employees in particular give you the benefit of the doubt. You know, you created this thing. And if you say, hey, we need to do a major shift in our strategy, even hard things like layoffs. Founders tend to get a lot of
Starting point is 00:16:45 latitude and are judged, I think, differently. And I think rightfully so in some ways because of the interconnection of their identity to the thing that they've created. And so I actually really believe in founder-led companies. One of the real interesting challenges is going from a founder-led company to not. And, you know, Amazon has gone through that transition. Microsoft has gone through that transition for that reason. But I love working with founders. And I, I love working with people like Toby and Sam because they're so different than me yet. And I can see how they operate their businesses and I am inspired by it. I learn from it. And obviously, working for market. Salesforce, you, I'm like, wow, that's really interesting. I'm almost like an anthropologist.
Starting point is 00:17:29 Why did you do that? You know, I want to learn more. And so I love working with founders that inspire me because I just learned so much from them. It's such an interesting front row seat into what's happening. Do you think founders go astray when they start listening to too many outside voices? And this goes back to the, I'm sure you're aware, the Brian Chesky, the founder mode. Do you think, talk to me about that. I have such a nuanced point of view on this because it is decidedly not simple. So broadly speaking, I really like the spirit of founder mode, which is just having.
Starting point is 00:18:04 having deep founder-led accountability for every decision at your company. I think that that's how great companies operate. And when you, you know, proverbially make decisions by committee or you're more focused on process than outcomes, that produces all the experiences we hate as employees, as customers. That's the proverbial DMV, right? You know, it's like process over outcomes. And then similarly, you look at the, disruption in all industries right now because of AI, you know, the companies that will recognize where things are clearly going to change. Like, everyone can see it. It's like a, you know, slow motion car wreck. Everyone knows how it ends. You need that kind of decisive breakthrough, boundaries, layers of management to actually make change as fast as required in business right now. The issue, I have not with Brian's statements, Brian's amazing, is how people can sort of interpret that and sort of execute it as a caricature
Starting point is 00:19:06 of what I think it means. You know, there was a, I remember after Steve Jobs passed away and, you know, I don't know, I've met Steve a couple times I haven't never worked with him in any meaningful way. You know, but he was sort of,
Starting point is 00:19:21 if you believe the story is like kind of pretty hard on his employees and very exacting. And I think a lot of founders were like mimicking that, you know, done to wearing a black turtlene, and yelling at their employees. I'm like, I'm not sure that was the cause. You know, I think Steve Jobs' taste and judgment
Starting point is 00:19:37 through, you know, executed through that, you know, packaging was the cause of their success. And somehow, and then similarly, I think founder mode can be weaponized as an excuse for just like overt micromanagement. And that probably won't lead to great outcomes either. And most great companies are filled with extremely great individual contributors who make good decisions and work really hard. and companies that are like solely executing through the judgment of individual probably aren't
Starting point is 00:20:08 going to be able to scale to be truly great companies. So I have a very nuanced point because I actually believe in founders. I believe in actually that accountability that comes from the top. I believe in cultures where, you know, founders have license to go in and all the way to a small decision and fix it, the infamous question mark emails from Jeff Bezos, you know, that type of thing. That's the right way to run a company. But that doesn't mean that you don't have a culture where individuals are accountable and empowered. And you don't want people trying to make business decisions because of what will please our dual leader, which is like the caricature of this. And so after that came out, I could sort of see it all happening, which is like some people will take that.
Starting point is 00:20:50 I'm like, you know what? You're right. I need to go down and be in the details. And some people will do it and probably make everyone who works for them miserable and probably both will happen as a consequence. totally thank you for the detail and nuance there i love that too do you think engineers make good leaders i do think engineers make good leaders but one thing i've seen is that i think that i really believe that great CEOs and great founders um start usually with one specialty but become uh more broadly specialists in our parts of their business um you know i think
Starting point is 00:21:28 I think the businesses are multifaceted and rarely as a business's success due to one thing, like engineering or product, which is where a lot of founders come from. Often your go-to-market model is important for consumer companies. How you engage with the world in public policy becomes extremely important. And I think as you see founders grow from doing one thing to growing to being a real meaningful company like Airbnb or meta or something, you can see those founders really transform from being one thing to many things. So I do think engineers make great leaders. I think the first principles thinking, the system design thinking, really benefits things like organization design, strategy. But I also think that, you know, when we were speaking earlier about identity, I think one of the main transitions founders need to make, especially engineers,
Starting point is 00:22:26 is you're not like the product manager for the company, you're the CEO. And at any given day, do you spend time recruiting an executive because you have a need? Do you spend time on sales because that will have the biggest impact? Do you spend time on public policy or regulation? Because if you don't, it will happen to you and could really impact your business in a negative way. And I think engineers who are unwilling to elevate their identity from what they were to what it needs to be in the moment often leads to sort of plateaus in companies growth. So 100 percent, I think engineers make great leaders. And it's not a coincidence, I think, that most of the Silicon Valley, great Silicon Valley CEOs came from engineering backgrounds.
Starting point is 00:23:17 But I also don't think that's sufficient either as your company scales. And I think that making that transition, as all the great ones have, is incredibly important. To what extent are all business problems, engineering problems? That's a deeper philosophical question that I think I have the capacity to answer. What is engineering? What I like about approaching problems as an engineer is first principles thinking and understanding the root causes of issues rather than simply addressing the symptoms of the problems. And I do think that coming from a background in engineering, that is everything from process, like how engineers do a root cause analysis of an outage on a server is a really great way to analyze why you lost a sales deal. You know, like I love the systematic approach of engineering.
Starting point is 00:24:07 One thing that I think going back to good ideas that can become caricatures of themselves, like one thing I've seen though with engineers who go into other disciplines is sometimes you can over. overanalyze decisions in some domains. Let's just take modern communications, which is driven in social media and very fast-paced. Having a systematic first principles discussion about every tweet you do is probably not a great comm strategy. And so, and then similarly, you know, there are some aspects of, say, enterprise software sales that, you know, aren't rational, but they're human, you know, like forming personal relationships, you know, and the importance of those to building trust with a partner. It's not all just, you know, product and technology. And so I would say I think a lot of things coming with an engineer mindset could really benefit, but I do think
Starting point is 00:25:07 that taking that to its, like, logical extreme can lead to analysis paralysis, can lead to over-intellectualizing some things that are fundamentally human problems. And so So, yeah, I think a lot can benefit from engineering, but I wouldn't say everything's an engineering problem in my experience. You brought up first principles a couple of times. You're running your third startup now, Sierra. It's going really well. How do you use first principles in terms of how do you use that at work? Yeah, it's particularly important right now because the market of AI is changing so rapidly. So if you rewind two years, you know, most people, hadn't used chat GPT yet. Most companies hadn't heard the phrase large language models or generative AI yet. And in two years, you have chat GPT becoming one of the most popular consumer services in history faster than any service in history. And you have across so many
Starting point is 00:26:11 domains in the enterprise, really rapid transformation. The law is being transformed. marketing is being transformed, customer service, which is where my company, Sierra Works is being transformed. Software engineering is being transformed. And the amount of change in such a short period of time is, I think, unprecedented. And perhaps I lack the historical context, but it feels faster than anything I've experienced in my career. And so as a consequence, I think, if you are responding to the facts in front of you and not thinking from first principles about why we're at this point and where it will probably be 12 months from now,
Starting point is 00:26:54 the likelihood that you'll make the right strategic decision is almost zero. So as an example, it's really interesting to me that with modern large language models, one of the careers that is being most transformed is software engineering. And, you know, one of the things I think a lot about is how many software engineers will we have our company three years from now? What will the role of a software engineer be as we go from being authors of code to operators of code generating machines? What does that mean for the type of people we should recruit? And if I look at the actual
Starting point is 00:27:33 craft of software engineering that we're doing right now, I think it's literally a fact that will be completely different to years from now. Yet I think a lot of people building companies hire for the problem in front of them rather than doing that. But two years is not that long. Those people that you hire now will just be getting really productive a couple years from now. So we try to think about most of our long-term business from First Principles, everything from, I'll say a couple examples in our business. Our pricing model is really unique and comes from First Principles thinking rather than having our customers pay a license for the privilege of using our platform. we only charge our customers for the outcomes, meaning if the AI agent they've built for their
Starting point is 00:28:17 customers solves the problem, there's usually a pre-negotiated rate for that. And that comes from the principle that in the age of AI, software isn't just helping you be more productive, but actually completing a task. What is the right and logical business model for something that completes a task? Well, charging for a job well done, rather than charging for the privileges using the software. Similarly, you know, we with a lot of our customers, you know, we help deliver them a fully working AI agent. We don't hand them a bunch of software and say, good luck, you know, configure it yourself. And the logic there is, you know, in a world where making software is easier than it ever is before
Starting point is 00:28:59 and you're delivering outcomes for your customer, the delivery model of software probably should change as well. And we've really tried to reimagine what, like, the software company of the future should look like and trying to, you know, model that in everything that we do. That's brilliant. How do you think software engineering will change? Is that you're going to have fewer people or the people are going to be organized differently? Or how do you see that? How geeky can I get?
Starting point is 00:29:26 As geeky as you want, man. I actually wrote a blog post right before Christmas about this. I think this is an area that deserves a lot more research. I'll describe where I think we are today, and smart people may disagree, but a lot of the modern large language models, both the traditional large language models and sort of the new reasoning models are trained on a lot of source code, and it's an important input to all of the knowledge that they're trained on. As a consequence, even the early models were very good at generating code. So, you know, every single engineer at Sira uses a cursor, which is a great product that basically integrates with the IDE Visual Studio code to help you generate code more quickly. It feels like a local maximum in a really obvious way to me, which is you have a bunch of code written by people, written in programming languages that were designed to make it easy for people. to tell a computer what to do.
Starting point is 00:30:32 Probably this funniest example of this is Python. It almost looks like natural language. Yeah. But it's notoriously not robust. You know, most Python bugs are found by running the program because there's not static type checking. Similarly, there's most bugs, while you could run a fancy static analysis, like most bugs show up simply at runtime because it's just not designed.
Starting point is 00:31:00 It's designed to be ergonomic to write. Yet, we're using AI to generate that. And so we've sort of designed most of our computer programming systems to make it easy for the author of code to type it quickly. And we're in a world, we're actually generating code is going to, like, the marginal cost of doing that is going to zero. But we're still generating code in programming languages that were designed for human authors. And similarly, if you've ever looked at someone else's code, which a lot of people do professionally is called a code review, it's actually quite hard to do a code review. You know, you end up interpreting, you're trying to basically put the system in your head
Starting point is 00:31:48 and simulate it as you're reading the code to find errors in it. So the irony now that have taken things that are code programming languages that were designed for authors, And now having humans do the job of essentially code reviewing, code written by an AI, and yet all of the AI being in the code generation part of it, I'm like, I'm not sure it's great, but we're generating a lot of code with similar flaws to that we've been generating before from security holes
Starting point is 00:32:16 to just functional bugs and in greater volumes. And I think what I would like to see is if you start with the premise that generating code is free or going towards free, what would be the programming systems that we would design? So, for example, Rust is an example of a programming language that was designed for safety, not for programming convenience. You know, my understanding is that the Mozilla project, you know, there were so many security holes in Firefox.
Starting point is 00:32:49 They said, let's make a programming language that's very fast, you know, but everything can be checked statically, including memory safety. Well, it's a really interesting direction where you weren't optimizing for authorship convenience or optimizing for correctness. Are there programming language designs that are designed so a human looking at it can very quickly evaluate, does this do what I intended it to do? There's an area of computer science I studied in college called formal verification, which at the time was turning a lot of computer programs into math proofs and finding inconsistencies, and it sort of worked well as not as well as you'd hope. But, you know, in a world where AI is generating a lot of code, you know, should we be investing more informal verification so that the operator of that code generator machine can more easily verify that it does, in fact, to do what they intended is to do, and could a combination of a programming language that is more structurally correct and structurally safe and exposes more primitives for verification plus a tool to verify, could you make an operator of a code generating machine? 20 times more productive, but more importantly, make the robustness of their output 20 times greater. And then similarly, you know, there's themes, things go in and out of fashion, but like test
Starting point is 00:34:06 driven development, you know, where you write your unit test first or your integration test first and then write code until it fulfills the test. Most programmers I know who are really good, not despise it, but it's just like a, it sounds better than it is in practice. But again, writing code is free, you know, so writing tests is free. You know, how can you create a programming system where the combination of great programming language design, formal verification, robust tests because you didn't have to do the tedious part of writing them all, could you make something that made it possible to write increasingly complex systems that were increasingly robust? And then similarly like the elephant in the room for me is the anchor tenant of most of these
Starting point is 00:34:46 code generating systems are an IDE right now, you know, and that obviously doesn't seem as important in this world. And even with coding agents, which is sort of where the world is going, it doesn't change the fact that, like, you know, who's accountable for the quality of it, who's fixing it. And I think there is a world where we can make reasonable software by just automating what we as software engineers do every day. But I have a strong suspicion that if we designed these systems with the role of a software engineer in mind being an operator of a machine rather than the author of the code, we could make the process much more robust and much more productive. And it feels like a research problem to me. It doesn't feel, and I think a lot
Starting point is 00:35:31 of people, and for a good reason, including me, are just excited about the efficiency of software development going up. And I want to see the new thing. I'm constructively dissatisfied with where we are. It's so interesting that if software AI is good enough to write the code, should we get enough check the code? That's a great question. But actually, you'll, you know, it's still funny to me that we'd be generating Python, you know, just because for anyone who's listening right now has ever operated a web service running Python, it's CPU and intensive, really inefficient, you know, should we be taking most of the unsafe C code that we've written and converting it to a safer system like Rust? You know, if authoring these things
Starting point is 00:36:12 and checking it are relatively free, shouldn't all of our programs be incredibly efficient? Should they all be formally verified? Should they all be? analyzed by a great agent. I do think it can be turtles all the way down. You can use AI to solve most problems in AI. The thing that I'm trying to figure out is like what is the system that a human operator is using to orchestrate all those tasks? And, you know, I go back to the history of software development. And most of the really interesting metaphors in software development came from breakthroughs in computing. So, you know, the C programming language came from Unix and when these time sharing systems were really, it went from sort of punch cards to
Starting point is 00:36:51 something that were a lot more agile. Small talk came out of the development of the graphical user interface at Xerox Park and, you know, there was a sort of a confluence of message passing as a metaphor and the graphical user interface. And then there was a lot of really interesting principles that came out of networking, you know, and sort of distributed systems, distributed locking, sequencing, I think we should recognize that we're in this brand new era as significant as the GUI. It's like a completely new era of software development. And if you were just to say, I'm going to design a programming system for this new world from first principles, what would it be? And I think when we develop it, I think it will be really exciting because
Starting point is 00:37:36 rather than automating and turning up the speed of just generating code and with the same processes we have today, I think will feel native to this system. and give a lot more control to the people who are orchestrating the system in a way that I think will really benefit software overall. Let's dive into AI a little bit. How would you define AGI to the layman? I think a reasonable definition of AGI might be that any task that a person can do at a computer, that system can do on par or better. I'm not sure that's a precise definition, but I'll tell you where that comes from and it's flaws, but there's not a perfect definition of AGI, in my opinion, or there's not a precise definition of AGI. I'm sure there's good answers.
Starting point is 00:38:30 One of the things about the G and AGI is about generalization. So can you have a system that is intelligent in domains that it wasn't explicitly trained to be intelligent on? And so I think that's one of the most important things is like given a net new domain, can this system become more competent and more intelligent than a person sort of trained in that domain? And I think that's sort of the, you know, at or better than a person is certainly a good standard there. And that's sort of the definition of superintelligence. the reason I mentioned at a computer is I do think that it is a bar that means like if there's a digital interface to that system, it affords the ability for AI to interact with it, which is why that's a bar that's reasonable to hit.
Starting point is 00:39:27 I say that because one of the interesting questions around AGI is how quickly it does generalize. And there are domains in the world that the progress in that domain isn't necessarily limited by intelligence, but by other social artifacts. So as an example, and I'm not an expert in this area, but if you think about the pharmaceutical industry, my understanding is, you know, the one of the main bottlenecks is clinical trials so no matter how
Starting point is 00:40:07 intelligent a system would be in discovering new therapies it may not materially change that and so you may have something that's discovering new insights in math and that would be
Starting point is 00:40:21 delightful and amazing but the existence of that system that's super intelligent in one domain may not translate to all domains equally. I just heard at least a snippet of a talk by Taller Cohen, the economist, and it was really interesting to hear his framing on this about which parts of the economy could sort of absorb intelligence more quickly than others.
Starting point is 00:40:46 And so I choose that definition of AGI, recognizing that there's not a perfect definition because it captures the ability of this intelligence to generalize while also recognizing that the domains of society like it might not apply with equal velocity even once we reach that point of a system being able to have that level of intelligence. You can get protein at home
Starting point is 00:41:09 or a protein latte at Tim's. No powders, no blenders, no shakers. Starting at 17 grams per medium latte, Tim's new protein lattes, protein without all the work, at participating restaurants in Canada. Summer's here and you can now get almost anything you need for your sunny day.
Starting point is 00:41:26 delivered with Uber Eats. What do we mean by almost? Well, you can't get a well-groom lawn delivered, but you can get a chicken parmesan delivered. A cabana? That's a no, but a banana, that's a yes. A nice tan, sorry, nope. But a box fan, happily yes. A day of sunshine? No.
Starting point is 00:41:41 A box of fine wines? Yes. Uber Eats can definitely get you that. Get almost, almost anything delivered with Uber Eats. Order now. Alcohol and select markets. Product availability may vary by Regency app for details. When I think about what artificial intelligence
Starting point is 00:41:56 is limited by or the bottlenecks if you will I keep coming back to a couple things there's regulation there's compute there's energy there's data and there's LMs
Starting point is 00:42:07 am I missing anything so you're saying the ingredients to AGI yeah like there's limitations on each aspect of those things and there seem to be the main contributors to what's limiting us from even accelerating
Starting point is 00:42:22 at this point is that how do you think about that? Yeah, what you said is roughly how I think about it. I'll put it into my own words, though. I think the three primary inputs are data, compute, and algorithms. And data is probably obvious, but one of the things after the Transformer model was introduced is it afforded an architecture with just much greater parallelism, which meant models could be much bigger and train more quickly on much more data, which just led to a lot of the breakthroughs with, that's the LLM, just they're large.
Starting point is 00:43:00 Yeah. And the scaling laws, you know, a couple of years ago, you know, indicated, like, the larger you make the model, the more intelligent would be at a degree of efficiency that was tolerable. And there, we are, you know, there's lots of stuff written about this, but, you know, there's, in terms of just, like, text. content to train on, you know, the availability of new content is certainly waning. And some people would say I think there's like a data wall. I'm not an expert in that domain, but it's been talked about a lot and you can read a lot about it. There's a lot of interesting opportunities though to generate data too. So there's a lot of people working on simulation. If you think about
Starting point is 00:43:42 a domain like self-driving cars, simulation is a really interesting way to generate. Is that synthetic data? Is that what? Yeah, I would say that synthetic data. The synthetic data has a simulation and synthetic data are a little different. So you can generate synthetic data, like generate a novel. Simulation, I would put at least in my head, and I'm sure that academics might critique what I'm saying. But I view simulation is based on a set of principles like the laws of physics. So if you're able to build a real world simulation for training a self-driving car, you're not just generating arbitrary data.
Starting point is 00:44:21 like the roads don't turn into loop-de-loops, you know, because that's not possible with physics. So by constraining a simulation with a set of real-world constraints, the data has more efficacy, you know? And so, and there's sort of a, it constrains the different permutations of data you can generate from it. So it's, I think, a little bit higher quality. But then along those lines, you know, a lot of people wonder if you generate synthetic data, How much value can that add to a training process? You know, is it sort of regurgitating information already had? What's really interesting about, you know, reasoning and reasoning models is I think I feel
Starting point is 00:45:04 really optimistic these models are generating net new ideas, and so it really affords the opportunity to break through some of these, the data wall as well. So data is one thing, and I think both synthetic data and simulation are really interesting opportunities to grow there. Then you have compute. And this is something that, you know, that's why there's so many data center investments. It's why NVIDIA as a company has grown so much.
Starting point is 00:45:35 Probably the more interesting kind of breakthroughs there are these reasoning models where there's not quite such a formal separation between the training process and the inference process where you can spend more compute at, the time of inference to generate more intelligence, which has really been a breakthrough in a variety of ways, I think is really interesting, but it shows you how you can run up against walls
Starting point is 00:45:57 and find new opportunities to use it. And then finally, algorithms. And the biggest breakthrough is obviously the Transformers model, attention is all you need, that paper from Google, that sort of led to where we are now. But there's been a number of really important papers since then from the idea of chain of thought reasoning to what at Open Eye, what we did with the O1 model,
Starting point is 00:46:17 which is to do some reinforcement or in those chains of thought to really reach new levels of intelligence. And so I do think that I mentioned some anecdotes about some breakthroughs there because my view is that each one of them has their own problems.
Starting point is 00:46:34 You know, compute. It's very capital intensive. And a lot of these models, the half-life of their value is pretty short because new ones come out so frequently. And so, you know, you wonder, like, you know, can we afford? What's the business case for investing this CapEx? And then you have a breakthrough like, you know, oh one and you're like, gosh, you know, with a distilled
Starting point is 00:46:58 model and moving more to inference time, it changes the economics of it. You have data. You say, gosh, we're running out of textual data to train on. Well, now we can generate reasoning. We can do simulations. Oh, that's an interesting breakthrough. And then on the algorithm side, as I mentioned, just the idea of these reasoning models is really novel itself. And each of these at any given point, if you talk to an expert in one of them and I'm an expert in none of them, they will tell you the sort of current plateau that they can see on the horizon. And there usually is one. I mean, you'll talk to different people about how long the scaling laws for something will continue and you'll get slightly different opinions, but no one thinks it's going to last forever.
Starting point is 00:47:36 And at each one of those, because you have so many smart people working on them, you often have people discovering a breakthrough in each of them. And so as a constant, I really do feel optimistic about the progress towards H.E.I. Because one of those plateaus might extend a while if we just don't have the key idea that we need to break through. The idea that we will be stuck on all three of those domains feels very unlikely to me. And in fact, what we've seen because of the potential economic benefits of AGIs, we're in fact seeing breakthroughs in all three of them. And as a consequence, you know, you're just seeing just the blistering pace of progress that we've seen over the past couple years.
Starting point is 00:48:18 At what point does AI start making AI better than we can make it, or making it better while we're sleeping? We can't be too far from that. Well, it might reflect back to our software engineering discussion, but broadly, this is the area of AGI around self-improvement, which is meaningful from an improvement standpoint, but also obviously from a safety standpoint as well. So, I don't know when that will happen, but I do think, you know, by some definition, you could argue that it's happening already in the sense that every engineer in Silicon Valley is already using coding agents and platforms like Cursor to help them code. So it's contributing already. And I imagine as coding assistants go to coding agents in the future, most engineers in Silicon Valley will show up in the morning. But this is sort of the different. between, you know, the assisted driving and Tesla versus, like, self-driving.
Starting point is 00:49:20 Yeah. Right? Like, at what point do we leap from, I'm a co-pilot in this to, I don't have to do anything? I mean, it's a question that there's so much nuance to answer. I'm not sure to answer because I'm not sure you'd want to necessarily. Like, I think for some software applications, that's important. But when we brought up, you know, we were talking about the active software development, people have to be accountable for the software that they produce.
Starting point is 00:49:46 And that means if you're doing something simple like a software-as-a-service application, that it's secure, that it's reliable, that the functionality works as intended for something as meaningful as an agent that is somewhat autonomous, does it have the appropriate guardrails, does it actually do what the operator's intended, is there appropriate safety measures. So I'm not sure there's really any system where you'd want to turn a switch and go get your coffee.
Starting point is 00:50:18 But I do think to the point on, you know, these broader safety things is I think that when you think about more advanced models, we need to be developing not only more and more advanced safety measures and safety harnesses, but also using AI to supervise AI and things like that. So it's a part, probably my colleague on the board Zico Coulter is probably a better person to talk through some of the technical
Starting point is 00:50:44 things, but there's a lot of prerequisites to get at that point, and I'm not sure it's simply like the availability of the technology, just because it is that at the end of the day, we are accountable for the safety of the systems we produce, not just opening out, like every engineer, and that's a principle that should not change. What does that mean, like when we say safety and AI, that seems so vague in general that everybody interprets it quite differently. Like, how do you think about that? And how do you think about that in the world where, let's say, we regulate safety in
Starting point is 00:51:18 the United States and another country doesn't regulate safety? How does that affect the dynamic of it? I'll answer broadly and then go into the regulatory question. So I really like Open AI's mission, which is to ensure that AGI benefits all of humanity. that isn't only about safety and I believe intentionally so though obviously the mission was created prior to my arrival
Starting point is 00:51:43 because it's both about safety kind of hypocritic of first do no harm and I don't think one could credibly achieve that mission if we created something unsafe so I would say that's the most important part of the mission but there's also a lot of other aspects of benefiting humanity is it universally accessible
Starting point is 00:52:00 is there a digital divide where some people have access to AGI and some day don't. Similarly, you could argue that does it, are we maximizing the benefits and minimizing the downsides? Clearly, AI will disrupt some job, but it also could democratize access to health care, education, expertise. So as I think about the mission, it starts with safety, but I actually like thinking about it more broadly because I think at the end of the day, benefiting humanity is the mission and safety is a prerequisite. But, safety is a prerequisite, but it's almost like going to my analogy of the Hippocratic Oath, a doctor's job is, you know, to cure you.
Starting point is 00:52:41 First, do no harm, but then to cure you. And a doctor that did no harm but didn't cure you wouldn't be great either. So I really like to think about the holistically. And again, Zika or Sam might have a more complete answer here. But broadly, I think about does the system that represents AGI align with the intentions of the people created it and the intentions that people operate? it so that it does what we want and it's a tool that benefits humanity, a tool that we're actively using to affect the outcomes that we're looking for. And that's kind of the way I think about safety. And it can be meaningful things like misalignment or more subtle things like
Starting point is 00:53:24 unintended consequences. And I think that latter part is probably the area that's really interesting from an intellectual and ethical standpoint as well. If I look at what was the bridge in Canada that fell down where it motivated the ring that a lot of engineers were. Oh, yeah, I forget the name of it. Just like whether it's the Tacoma Narrows Bridge in Washington or Three Mile Island or these intersections where we've engineered these, you know, what at the time people hope would be positively impact humanity, but something went horribly wrong. Sometimes it's engineering, sometimes it's bureaucracy, sometimes it's a lot of things. And so I don't think, when I think about safety, I don't just look at the technical measures of it, but how does this technology manifest in society? How do we make decisions around it?
Starting point is 00:54:18 And you could take, put another way, technology is rarely innately good or bad. It's sort of what we do with it. And I think those social constructs matter a lot as well. So I think it's a little early to tell because we don't have this kind of superintelligence right now. And I think it won't just be a technology company defining how it manifests in society. And you could imagine taking a very well-aligned AI system and a human operator directing it towards something that would objectively hurt society. And there's a question of like, who gets to decide, who's accountable? And it's a perennial question.
Starting point is 00:54:56 I mean, it's whether you're deciding, you know, should you use your. smartphone in school, you know, who should decide that. And I, there's parents who will tell you, hey, it's my decision. It's my kid. And then there's principals who will tell you it's not benefiting the school. And I'm not sure that's going to be my place or our place, but there'll be a number of those conversations that are much deeper than that question that I think we'll need to answer. As it relates to regulation, there's two not conflicting forces, but two forces that exist somewhat independently, but relate to each other. One is the pace of the progress in AI and ensuring that, you know, the folks working on frontier models are ensuring
Starting point is 00:55:39 those models do benefit humanity. And then there's the sort of geopolitical landscape, which is, you know, do you want AGI to be created by the freedom, sort of the West, by democracies, or do you want it to be created by more totalitarian governments? And so I think the inherent tension for regulators will be a sense of obligation to ensure that, you know, the technology organizations creating AGI are, in fact, focusing enough on that infity and humanity, all the other stakeholders who's interests that they're accountable for and ensuring that the West remains competitive. And I think that's a really nuanced thing. And I think, you know, my view is it's very important that the West leads in AI, and I'm very proud of the fact
Starting point is 00:56:37 that, you know, Open AI is based here in the United States, and we're investing a lot of the United States. And I think that's very important. And I also, you know, having sort of seen the inside of, I think we're really focused on benefiting humanity. So I tend to think that, you know, it needs to be a multi-stakeholder dialogue, but I think there's a really big risk that some regulations could have the unintended consequence of slowing down this larger. conversation. But I don't say that to be dismissive of it either. It's actually just an impossibly hard problem. And I think you're seeing it play out, as you said, in really different ways. In Canada, United States, Europe, China, elsewhere. I want to come back to
Starting point is 00:57:14 compute and the dollars involved. So, I mean, on one hand, you have, if I could start an AI company today by, you know, putting my credit card down using AWS and leveraging their infrastructure, which they've built, they've spent the hundreds of billions of dollars. and I get to use it on a time-based model. On the other hand, you have people like OpenAI and Microsoft investing tons of money into it that may be more proprietary. How do you think about the different models competing?
Starting point is 00:57:48 And then the one that really throws me for a bit of a loop is Facebook. So Facebook has spent... You know the matter made a new. Oh, God. So meta, I'm like aging myself here. So meta comes along and, you know, possibly for the good of humanity, but, like, I tend to think Zuck is, like, incredibly smart. So I don't think he's spending, you know, $100 billion to develop a free model and give it away to society. How do you think about that in terms of return on capital and return on investment?
Starting point is 00:58:22 It's a really complicated business to be on just given the CAPEX required to build a frontier model. But let me just start with a couple definitions of terms that I think are useful. I think most large language models I would call foundation models. And I like the word foundation because I think it will be foundational to most intelligent systems going forward. And most people building modern models, particularly if they involve language, image or audio, shouldn't start from building a model from scratch. They should pick a foundation model, either use it off the shelf. or fine-tune it. And so it's truly foundational in many ways.
Starting point is 00:59:01 In the same way, most people don't build their own servers anymore. They lease them from one of the cloud infrastructure providers. I think foundation models will be something trained by companies that have a lot of CAPEX and leased by a broad range of customers who have a broad range of use cases. And I think that leads in the same way that data center builders having a lot of data centers enabled you to have the capital scale to build more data centers. I think the same will largely be true of, you know, building the huge clusters to do training and things like that. Foundation models, I think, are somewhat distinct from frontier models. And frontier
Starting point is 00:59:40 models, I think it's a term credit to Reed Hoffman, but I may be mistaken on that, but that's where I heard it from. And these are the models that are usually like the one or two that are clearly the leading edge, O3, as an example, from Open AI. And these frontier models, are being built by labs who are trying to build AGI that benefits humanity. And I think if you're deciding whether you're building a foundation model and what your business models around it, it's very different business, then I'm going to go pursue AGI. Because if you're pursuing AGI, really, there's only one answer,
Starting point is 01:00:20 which is to build and train and move to the next horizon, because if you can truly build something that is AGI, the economic value is so great, I think there's a really clear business case there. If you're pre-training a foundation model that's the fourth best, that's going to cost you a lot of money. And the return on that investment is probably fairly questionable because why use your fourth best large language model versus a frontier model or an open source one for meta. And as a consequence of that, I think we have probably have too many people building models
Starting point is 01:00:59 right now. There's already been some consolidation, actually, of companies being folded into Amazon and Microsoft and others. But I do think it will play out a bit like the cloud infrastructure business where a very small number of companies with very large KAPX budgets are responsible for both building and operating these data centers. And then developers and in consumers will use things like chat GPT as a consumer or as a developer, you'll license and rent, you know, one of these models in the cloud. How it will play out is a really great question. You know, I think the, I heard one investor talk about these as like the fastest depreciating
Starting point is 01:01:39 assets of all time. On the other hand, you know, I, if you look at the revenue scale of something like an open AI and what I've read about, places like Anthropic, let alone Microsoft and Amazon, it's pretty incredible as well. And so you can't really, if you're one of those firms, you can't afford to sit on the sidelines as the world transforms. But I would have a hard time personally, like, funding a startup that says I'm going to do pre-training.
Starting point is 01:02:08 You know, I don't really know, like, what's your differentiation in this marketplace? And I think a lot of those companies, you're already seeing them consolidate because they have the cost structure of a pharmaceutical company, but not the business model. This is just it, though, right? Like, Open AI has a revenue model around a revenue model. Microsoft has a revenue model around their AI investments. They just updated the price of teams with co-pilot. Amazon has a revenue model around AI, in a sense,
Starting point is 01:02:41 they're getting other people to pay for it through AWS, and then they're getting the advantages of it at Amazon, too, from a consumer point of view. And all the millions of projects, Bezos was doing an interview last week. that there's every project at Amazon basically has an AI component to it now. Facebook, on the other hand, has spent all of this money already and with an endless amount, presumably in sight, or like not in sight, an endless amount to go,
Starting point is 01:03:08 but they don't have a revenue model specifically around AI, where it would have been cheaper, obviously, for them to use a different model, but that would have required presumably giving data away. I'm just trying to work through it from Zuck's point, of you if you know i actually will take market as word and you know that post you wrote about open source i think was very well written and yeah totally people to read it i think that's a strategy and you know if you look at facebook um you know you've got me saying facebook too so that was that was what i messed you up yeah you know the company is always really embraced open source
Starting point is 01:03:44 and if i look at really popular things from react to uh you know now the llama models it's all has been a big part of their strategy to court developers around sort of their ecosystem. And Mark articulated some of the strategy there. And I'm sure there's elements of commoditizing your complement. But I also think that, you know, if you can attract developers towards models, there's a strength. I, you know, I'm not really on the inside there. So I don't really have a perspective on it other than I actually think it's really great that there's different players with different incentives all investing so much. And I think it is really furthering the cause of bringing these amazing tools to society. But a lot changes. I mean, if you look at the
Starting point is 01:04:31 price of GPT40 Mini, you know, it is so much higher quality than like the highest quality model two years ago and much cheaper. I haven't done the math on it, but it's probably cheaper to use that than to self-host any of the open source models. So even the existence of the open source models, it's not free. I mean, inference costs money. And so there's a lot of complexity here. And actually, I have the email even being relatively close to stuff. Like, I have no idea where things are going.
Starting point is 01:05:03 But, you know, you could talk to a smart engineer and they'll tell you, oh, yeah, if you built your own servers, you'll spend less than renting them from, say, Amazon web services. or Azure. That's sort of true in absolute terms, but this is the fact, like, do you want someone in your team building servers? Oh, and in fact, if you change the way your service works and you need a different SKU like you all of a sudden are doing training and you need Nvidia, you know, H-100s, now all of a sudden you're built servers like this, you know, asset that's worthless. So I think with a lot of these models, you know, the presence of
Starting point is 01:05:40 open source is incredibly important and I really appreciate it. I also think that the think like the economics of they are pretty complex because the hardware is very unique. The cost to serve is much higher. Techniques like distillation have really changed the economics of models whether or not it's open source or hosted and leased. So it's, I think broadly speaking for developers, it's kind of an amazing time right now because you have a like a menu of options that's incredibly wide. And I actually think of it as, you know, just like in cloud computing, you'll end up with a price performance quality tradeoff. And for any giving engineering talents, they'll have a different answer. And that's appropriate. And some people use open source Kafka. Some people
Starting point is 01:06:28 work with confluent. Great. You know, like that's just the way these things work, you know. So you don't think AGI is going to be like a winner take all? You think there's going to be multiple options that have by definition, whatever the definition is of AGI? Well, first, I think Open AI, I believe, will play a huge part in it because there's both the technology, which I think Open AI continues to lead on, but also Chat-G-T, which has become synonymous with AI for most consumers. But more than that, it is the way most people access AI today. And so one of the interesting things like, what is AGI? We talked about, you know, opinions on what the definition might be. But the other question is, like, how do you use that?
Starting point is 01:07:14 Like, what are you, what is the packaging? And some of intelligence will be simply the outcomes of it, like a discovery of a new drug, which would be, you know, remarkable and hopefully we can cure some illnesses. But others will be just how you as an individual access it. And, you know, most of the people I know, like if they're signing an apartment lease, we'll put it in a chat, GPD, you get a legal opinion. And if you get, you know, lab results from your doctor, you can get a second opinion on chat GPT. Clay and I use the O1 Pro mode for, like, criticizing our strategy at Sierra all the time.
Starting point is 01:07:55 And so for me, what's so remarkable about chat GPT, which was this, you know, quirkily named research preview that has come to be synonymous with AI as I do think that it will be the delivery mechanism for AGI. when it's produced, and not just because of the many researchers at OpenEI, but because of the amazing, like, utility it's become from individuals. And I think that's really neat because, I don't know if it would have been obvious. If we were having this conversation three years ago, you know, and you were talking about artificial general intelligence, I'm not sure either of us would have envisioned something so simple as a form factor to absorb it that you just talk to it.
Starting point is 01:08:37 So I think it's great, and especially as I think about, the mission of opening eye, which is to ensure that AGI benefits humanity. What a simple, accessible form factor. There's free tiers of it. Like, what a kick-ass way to benefit humanity. So I really think that will be central to what we come as society to define as a AGI. You mentioned using it at Sierra to critique your business strategy. What do you know about prompting that other people miss? I mean, you must have the best prompts. People think that, you know, because I'm affiliated with it. You're not going, like, here's my strategy.
Starting point is 01:09:13 What do you think? What are you putting in there? I often, with the reasoning models, which are slower, will use a faster model first, GBT40, to refine my prompts. So over the holidays, partly because I was thinking about the future of software engineering, I've written a lot of compilers in my time. I'm, like, written enough that I, you know, it's like, a. It's easy for me. So I decided to see if I could have 01 Pro mode, generate end-to-end, a compiler, front-end, parsing, the grammar, checking for semantic correctness, generating an intermediate representation, and then using LVM, which is sort of a compiler collection that's very popular to actually do, you know, run, run it all.
Starting point is 01:10:07 And I would spend a lot of time iterating on 4-0 to sort of like refine and make more complete and specific what I was looking for. And then I would put it into O1 Pro mode, go get my coffee and, you know, come back and get it. I'm not sure if that's a viable technique, but it's really interesting because I do think in the spirit of AI being the solution to more problems than AI, having a lower latency, simpler model, help refine. essentially, I like to think of it as like you're like a product manager and you're asking an engineer what to do is your product requirements document complete and specific enough
Starting point is 01:10:45 and waiting for it is sometimes slower than and so I like doing it in stages like that. So that's my chip. At some point there's probably someone from open AI listening is going to like roll their eyes but that's just that's I. Who can I talk to at Open AI that's like the prompt ninja? I'm like so curious about this because I've
Starting point is 01:11:01 actually taken recently to getting open AI or chat TBT I guess if you want to call it I've been getting
Starting point is 01:11:10 chat GBT to write the prompt for me so I'll prompt it with I'm prompting an AI here are the key things
Starting point is 01:11:17 similar to my technique I want to accomplish what would an excellent prompt look like and then I'll copy paste that prompt that it gives me
Starting point is 01:11:26 back into the system but I'm like I wonder what I'm missing here right it's a good technique I mean there's lots of engineering techniques like that, like self-reflection is a technique where you have a model, observe, and critique, you know,
Starting point is 01:11:39 a decision like a chain of thought. So in general, you know, that mechanism of self-reflection is, I think, a really effective technique. You know, at Sierra, we help companies build customer-facing AI agents. So if you're setting up a Sona speaker, you'll now chat with an AI. If you're a serious XM subscriber, you can chat with Harmony, who's their AI to manage your account. We use all these tricks, you know, self-reflection to detect things like hallucination or decision-making, generating chains of thought for more complex tasks to ensure that it's, you know, you're putting as much compute and cognitive load into important tricks. So, you know, we're the, there's a whole industry around sort of figuring out how do you exact the like robustness and precision out of these models. So it's really fun, but changing rapidly. Hypothetical question, you've been hired to lead or advise a country that wants to become an AI superpower.
Starting point is 01:12:43 What sort of steps would you take? What sort of policies would you think would help create that? How would you bring investment from all over the world into that country? And researchers, right? So now all of a sudden you're competing. It's not the United States. How do you sort of set up a country from first principles all the way back to like, what does that look like? What are the key variables? Well, I mean, especially, this is definitely
Starting point is 01:13:07 outside of my domain of expertise, but I would say one of the key ingredients to modern AI is compute, which is a noun that wasn't a noun until recently, but now compute is a noun. And you know, I do think that's one area where policymakers can, because it involves a lot of things that touch federal and local governments like power land, and then similarly attracting the capital, which is immense to finance the real estate, to purchase the compute itself, and then to sort of operate the data center. And again, there's really immense power requirements for these data centers as well. And then, you know, it's attracting sort of the right researchers and research labs to, you know, leverage that, but in general, where there is compute, the research
Starting point is 01:14:04 labs will find you, you know, and I think that's it. And then there's a lot of national security implications, too, just because, you know, these models are very sensitive. At least the frontier models are. And so, you know, how you, your place in the geopolitical landscape is quite important. Like, will research labs and will the U.S. government be comfortable with training happening there and export restrictions and things like that. But I think a lot of it comes down to infrastructure as it relates to policy, is my intuition. I think right now so much of AI is constrained on infrastructure that that is the input to a lot of this stuff. And then there's a lot around attracting talent and all that. But as I said, you know, you look at the research labs. It's not
Starting point is 01:14:54 that many people, actually. It's a lot, but the compute is a limited resource right now. That's a really good way to think about it. I think about this for the lens of Canada, right? Which is like we don't have enough going on in AI. We tend to lose most of our great people to the states who then go to set up infrastructure here for whatever reason and don't bring it back to Canada. And I wonder how Canada can compete better. So this is like sort of the lens. I, like look at these questions through. How do you see that the next generation of education? Like if you were setting up a school today from scratch and again, hypothetical, not your
Starting point is 01:15:35 domain of expertise, but like using your lens on AI, how do you think about this? So like what skills will kids need in the future and what skills do we probably don't need to teach them anymore that we have been teaching them? Well, I'll start with the benefits that I think are probably obvious, but I'm incredibly excited about. I think education can become much more personalized. Oh, totally. Have you seen synthesis tutor, by the way? No, I have not. Oh, so they developed this synthesis, this AI company, developed this tutor, which actually teaches kids. And it's so good that El Salvador, the country, just recently adopted and replaced their teachers. And, like, it'll teach you, but it teaches
Starting point is 01:16:19 you specific to what you're missing. So it's not like every lesson's the same. It's, it's like, well, you're not understanding this foundational concept. So it's like K through five or six right now. That's amazing. And the results are like off the charts. Well, it doesn't surprise me. And I don't actually view it as like necessarily replacing a teacher. But my view is if you have a teacher with 28 kids in his or her class,
Starting point is 01:16:42 the likelihood that they all learn the same way or learn at the same pace is very unlikely. And, you know, I can really think of a, say, an English teacher or history teacher, you're orchestrating their learning journeys through a topic, say AP European history in the United States, there's a curriculum, they need to learn it, how someone will remember something or understand the significance of Martin Luther, you know, is very different. And you can, you know, generate an audio podcast for someone who might be an auditory learner. You can create cue cards for someone who needs that kind of repetition, repetition. You can visualize key moments in history for people who just maybe want to more viscerally
Starting point is 01:17:29 appreciate why this was a meaningful event rather than this dry piece of history. And all of that, as you said, can be personalized to the way you learn and how you learn. I think it's just incredibly powerful. And so one of the things I think is neat about AI is it's democratizing access to a lot of things that used to be fairly exclusive. A lot of wealthy people, if their child was having trouble in school, would pay for a tutor, math tutor, science tutor. And, you know, if you look at kids who are trying to get into, you know, big-name colleges, you know, if you have the means, you'll have someone prep you for the SATs or help you with your college essays, all of that should be democratized if we're doing our jobs well.
Starting point is 01:18:11 And it means that we're not living people's opportunity by their means. And I think that's just the most American thing ever, Canadian as well. It's the most incredible thing for humanity. It's the most incredible thing, humanity. And so I just think education will change for the positive in so many ways because I actually with my kids walking around when they ask, you know, if you have little kids, they ask why, why, why, and, you know, there's some point a parent just starts making up the answer or being dismissive. And like, we have chat to PT out. It's like the best when you're traveling and put on advance. voice mode and be like, ask away.
Starting point is 01:18:48 100%. And I'm listening too. You know, it's like you're, you live through your children's curiosity. And my daughter went to high school and came home with Shakespeare for the first time. And I was, she asked me a question. I was like, I felt this is like total inadequacy. I was like, I was very bad at this the first time. And then we put it into chat, GPT. And it was the most thoughtful answer. And she could ask follow up questions. And I actually was, you know, with her because I was like, oh, I forgot about that. You know, didn't even think about that. So I just think it's incredible, and I would like to, in public school systems, I think it's really, I think it'll be a really great when public school systems formally adopt these things so that they lean into tools like chat TPT as mechanisms to, like, raise the performance level of their classroom. And hopefully you'll see it in things like test scores and other things because kids can get the
Starting point is 01:19:46 extra time, even if the school system can't afford it for everyone. And then most importantly, kids care getting explanations according to their style of learning, which I think will be quite important as well. As it relates to skills, it's really hard to predict right now. And I would say that I do think learning how to learn and learning how to think will continue to be important. So I think most of, you know, primary and secondary education shouldn't and is not vocational necessarily. Some of it is, you know, I took auto shop and all of that, and I'm glad I did.
Starting point is 01:20:19 But I couldn't fix my electric car today with that knowledge. You know, things change. And I don't think it needs to be purely, you know, non-vocational. But, you know, the basics of learning how to think, learning, writing, reading, math, physics. chemistry, biology, not because you need to memorize it, but understand the mechanisms that create the world that we live in is quite important. I do think that there's a risk of people sort of becoming ossified in the tools that they use. So, you know, let's go back to our discussion of software engineering for a second, but I'll give other examples. You know, if you define your
Starting point is 01:21:07 as a software engineer is how quickly you type into your IDE, the next few years might leave you behind, you know, because that is no longer a differentiated, you know, part of the software engineering experience or it will not be. But your judgment as a software engineer will continue to be incredibly important in your agency and making a decision about what to build, how to build it, how to architect it, maybe using AI models as a creative foil. And so I think that just in the same way, if you're an accountant, using Excel doesn't make you less of an accountant. And just because you didn't, you know, hand craft that math equation, it doesn't make the results any less valuable to your clients. And so I think we're
Starting point is 01:21:56 going to go through this transformation where I think the tools that we use to create value in the world will change dramatically. And I think some people who define their jobs by their ability to use the last generation's tools really, really effectively will be disrupted. But I think if we can empower people and to re-skill and also broaden the aperture by which they define the value they're providing to the world, I think a lot of people can make the transition. The thing that is sort of uncomfortable, not really an education where it's just earlier in most people's lives. It's just I think the pace of change exceeds that of most technology transitions. And I think it's unreasonable to expect most people to change the way
Starting point is 01:22:45 to work that quickly. And so I think the next five years, I think will be, you know, for some jobs really disruptive and tumultuous. But if you take the longer view and you fast forward 25 or 50 years, I'm incredibly optimistic. I think it's the change will require. from society, from companies, and from individuals, like an open-mindedness about re-skilling and reimagining their job to the lens of this dramatically different new technology. At what point do we get to...
Starting point is 01:23:16 I mean, we're probably on the cusp of it now and it's happening in pockets, but what point do we start solving problems that humans haven't been able to solve or eliminating paths that we're on, maybe with medical research, that it's like, no, this whole thing, you've spent $30 billion on, you know, based on this
Starting point is 01:23:34 1972 study that was fabricated, but that one study had all these derivative studies and like, I'm telling you it's false, you know, because I can look at it through an objective lens and get rid of these 30 billion. You're smiling, so. Oh, no, I just, I hope soon. I mean, I hope, I mean, I, there was a lot of, there's a, one of the models I can't remember which one introduced a very long context window,
Starting point is 01:23:58 and there's a lot of people on X over the week. and putting in their thesis, you know, like grad school thesis in there. And it was actually critiquing them with, like, surprising levels of fidelity. And I think we're sort of there, perhaps, with the right tools, but certainly over the next few years. You know, we talked about what does it mean to generalize AI? Certainly in the areas of science that are, you know, largely represented through text and digital technological math being probably the most applicable. There's not really anything keeping AI from getting really good at math. There's not really an interface to the real world. You don't need to
Starting point is 01:24:40 do a clinical trial to verify something's correct. So I feel a ton of optimism there. It'll be really interesting in like, you know, areas of like theoretical physics. You'll continue to have the divide between the applied and the theoretical people. But I think there could be like really interesting new ideas there and perhaps some finding logical inconsistencies with some of the fashionable theories which has happened many times over the past few decades
Starting point is 01:25:08 I think we'll get there soon and actually what's really neat about is most of the scientists I know people who are actually doing science like they're the most excited about these technologies and they're using them already and I think that's really neat and I think we're hopefully going to be
Starting point is 01:25:24 I really hope we see more breakthroughs in science one of the things i have not an expert in but i've read a lot like a lot as a amateur about is just the slowdown and scientific breakthroughs over the past you know a few decades and and some theories that it's because of the degree of specialization that we demand of grad students and things like that and i hope with you know in general with i um democratizing access to expertise. I have a completely personal theory that it will benefit deep generalists in a lot of ways, too, because your ability to understand a fair amount in a lot of domains and leveraging AI,
Starting point is 01:26:09 knowing where to prompt the AI to go explore and bringing together those domains, it will start to shift sort of the intellectual power from people who are extremely deep to people who actually can orchestrate intelligence between lots of different domains or breakthroughs. I think that'll be really good for society because most scientific breakthroughs aren't. They tend to be cross-pollinating very important ideas from a lot of different domains, which I think will be really exciting. How important is the context window? I think it could be quite important, especially it certainly simplifies working with an AI. You can just give it everything and instruct it to do something.
Starting point is 01:26:52 And so, and assuming it works, you know, you can extend a context window and it can, the tension can be spread fairly thin and the robustness of the answer can be questionable. So, but assuming, unless just for argument's sake, you know, perfect robustness, I think it can really simplify the interface to AI. The not all uses, I also think that we're talking about open source models and APIs. I also think that most, what I'm excited about in the software industry is not necessarily a large language model with a prompt and a response being the product of AI, but actually end-to-end, close-loose systems that use large language models as pieces of infrastructure. And I actually think that a lot of the value in software will be that. And for many of those applications, the context window size can matter, but often because you have contextual awareness of the process that you're executing, context window is a little bit less important. So I think it matters a lot to intelligence. You know, there's a, I can't remember someone, one of the, some researcher said, you know,
Starting point is 01:28:03 you put all of human knowledge in the context window and you ask it to invent the next thing. You know, it's obviously a reductive thought, but interesting. But I actually, I'm equally excited about sort of the industrial applications of large language models, sort of like my company, Sierra. And if you're returning a pair of shoes at a retailer and it's a process that's, fairly complicated and, you know, is it within the return window? You know, do you want to return it in store? Do you want to send it? Do you want to print a QR code? Blah, blah, blah, blah. The orchestration of that is as significant as the models themselves. And I actually think as
Starting point is 01:28:38 we, just like computers, you know, there's going to be a lot of things where computers are a part of the experience, but it's not like manifesting itself as a computer. So I'm actually equally excited about those. And I think context window is slightly less important than those applications. Do you think that the output from AI should be copyrightable or patentable or – let me just take an example. If I go to the U.S. Patent Office, I download a patent for, let's say, the AeroPress, and I upload it to 01 Pro, and I say – I can't upload it yet because you don't let me do the P.D.S., but I upload it to four. And so I say, hey, what's the next logical leap that I could patent office? It would give me back diagrams and an output. And presumably, if I look at that and I'm like, yeah, that's legit.
Starting point is 01:29:25 I want to file that patent, can I? I don't know to answer that question. I'm not an expert in sort of intellectual property, but I think there will be an interesting question of, was that your idea because you used a tool to do it? I think the answer is probably yes, that you used the tool to do it. But I also think that the, in general, like the sort of marginal cost of intelligence will go down a lot. So a lot of the, you know, I think in general, like, will be in this renaissance of new ideas and intelligence being produced. And so I think that's broadly a good thing.
Starting point is 01:30:05 And I think, you know, the marginal value of that insight that you had might be lower than it was, you know, and years ago. What I was hoping you would say is that, you know, that's going to become less and less important because I feel like the patent trolls and all of the stuff that slows down innovation in some ways. Obviously, like, there's legitimate patents that people infringe on and there should be legal recourse. But if I could just go and patent like a hundred things a day, because it seems like that should not be allowed. This is what I'm saying, though. Well, in general, I think that, you know, companies, you know, I think patents make sense if it's predicting something that's at an active, use that you invented and you're you're trying to uh you know like the standard you know uh legal rationale for patents just generating a bunch of ideas and patented it seems destructive to the value
Starting point is 01:30:54 so here's the idea i had last night to counter this because i was like i don't want somebody doing this uh and i was thinking like what if prior art eliminates patents yeah so i was like what if i just set off like an instance and just publish it on a website nobody has to read that website here's a billion ideas. Exactly, but it's like basically patenting like anything, not patenting, but it's creating prior art for everything. So like you can't compete on that anymore. I don't know. I was like thinking about that. I thought it was fun. Reading, playing, learning. Stellist lenses do more than just correct your child's vision. They slow down the progression of myopia. So your child can continue to discover all
Starting point is 01:31:35 the world has to offer through their own eyes. Light the path to a brighter future, with Stellist lenses for myopia control. Learn more at SLR.com and ask your family eye care professional for SLR Stellist lenses at your child's next visit. At Grey Goose, we believe that pleasure is a necessity. That's why we craft the world's number one premium vodka in France,
Starting point is 01:32:00 using only three of the finest natural ingredients, French winter wheat, water from Jean-Sacke and yeast. With Grey Goose, we invite you to live in the moment and make time wait. Seep responsibly. Tell me about the Google Maps story. This is like now legend, and I want to hear it from you. This is my weekend coding.
Starting point is 01:32:25 Is that what you want to hear about? Yeah. Yeah, so I'll start with just like the story of Google Maps, the abbreviated version. We had launched a product at Google called Google Local, which was sort of a yellow pages search engine, you probably, probably most listeners don't even know what yellow pages are, but it was a thing back then. And we had licensed maps from MapQuest, which was the dominant sort of mapping provider at the time. And it was sort of an eyesore on the experience. And also always felt like it could be a more meaningful part of the kind of local search and
Starting point is 01:33:01 navigation experience on Google. So Larry Page in particular was really pushing us to really invest more to maps. We found this small. company with like four people in it, if I'm remembering correctly, started by Lars and Jens Rasmuson called Where2 Technologies, where they had made a Windows application called Expedition that was just a beautiful mapping product. It was running on Windows long after it was sort of out of fashion to make Windows apps, but they were sort of where the technology they're comfortable with. But they're really, their maps modeled the A to Z maps in the U.K. and were just beautiful, and they just had a lot of passion for mapping.
Starting point is 01:33:42 So we did a little aqua hire of them and took together the Google Local team and Lars and Jens's team and said, okay, let's take the good ideas from this Windows app and the good ideas from Google Local, and like, let's bring them together to make something completely new. And that's what became Google Maps. But there's a couple of idiosyncrasies in the integration. Because it was a Windows app, it really helped and hurt us. in a number of ways.
Starting point is 01:34:09 One of the ways it helped us is the reason why Google Maps we were able to drag the map and it was so much more interactive than any web application that preceded it was the standard that we needed to hit from interactivity was set by a native Windows app, not set by the legacy websites
Starting point is 01:34:28 that we had used at the time. And I think that by having the goalposts so far down the field because they had just started with this Windows app, which was sort of a quirk of Lars and Jens, just like technical choices, we made much bolder technical bets than we would have otherwise. I think we would have ended up much less interactive had we not started
Starting point is 01:34:50 with that quirky technical sort of decision. But the other thing was this Windows app, there's a lot of, like, it's hard to describe the like early 2000s of people who didn't live it, but like XML was like really in fashion. So like most things and Windows and other places was like XML and XSLT, which was a way of transforming XML into different XML was the basis of everything.
Starting point is 01:35:11 It was like all of enterprise software was like XML this, XML that. So similarly, when we were taking some of these ideas and putting them in a web browser, we kind of like went into autopilot and used like a ton of XML. And it made everything just like really, really tedious. And so Google Maps launched
Starting point is 01:35:30 with some really great ideas like the dragable maps. And we did a bunch of stuff for the local search technologies. you could overlay restaurant listings. It was really great. It was a really successful launch. We were like the hot shots within Google afterwards. But it really started to show its craft.
Starting point is 01:35:47 And we got to this point where we decided we wanted to support the Safari web browser, which was relatively new at the time. This was before mobile phones. And there was much less XML support in Safari than there was an Internet Explorer in Firefox. And so one of the engineers implemented like a full XSLT transform engine. in JavaScript to get it to work. And it was just like shit on top of shit, on top of shit. And so what was a really elegant fast web application had sort of quickly become something,
Starting point is 01:36:19 you know, there's a lot of dial-up modems at the time and other things. So like you'd show up to maps and it just was slow and like it just bothered me as like someone who takes a lot of pride in their craft. And so I got really energized and like over a weekend and a lot of coffee like rewrote. it um but it rewrote the whole thing there rewrote yeah more or less the whole thing and it took probably another week of like you know working through the bugs but yeah i sent it out to the you know the team after that weekend and it was it was the reason i was able to do it yeah i'm like a decent programmer but you know you'd also like lived with every bad decision up to that point too so i knew
Starting point is 01:37:00 exactly the output i was going to like i had simulated in my head like if i could do it over again and this is the way I'd do it. So by the time I, like, put my hands on the keyboard on, like, you know, Friday night, it wasn't like I was designing a product. Like, I knew I had been in every detail of that product since the beginning, including me the bad decisions, too, the not all the bad decisions. And so it was just very clear. I knew what I wanted to accomplish.
Starting point is 01:37:25 And for any, you know, any engineers worked on a big system, you have the whole system mapped out in your head. So I knew everything. And I also knew that, you know, there's a lot of pride of authorship with engineering and code so I sort of knew I really wanted to finish it over the weekend so that people could use it and see how fast it was and kind of overcome anyone who was like you know you know protective of the code they had written a few months ago and so I really wanted the prototype to go out so I did it and then I didn't it's funny I never talked about it again but I think Paul Bouhite who's was the co-creator of Gmail and and I worked and started friend feed with me he was on an interview and mentioned this story so now all of a sudden it's like everyone's talking about it and I was like well thank you Paul it's a little embarrassed so that people know about it that it was it's a true story and
Starting point is 01:38:16 and and XML is just the worst so did you get a lot of flack from the people who had built the system you effectively replaced like you were part of that team but everybody else had so much invested in it even though it was like shit on top of shit on top of shit you know I wrote a lot of it too so yeah I'm sure there was some around it but actually, I think good teams want to do great work. And so I think there was a lot of people constructively dissatisfied with the state of things too. And, you know, I think, you know, the engineer had written that XSLT transform, I think, was like, you know, a little bit. That's a lot of work. So you have to throw out a lot of work, which feels bad. But particularly, you know,
Starting point is 01:38:58 Lars and Yens and I, like, we want to make great products. And so I don't think there was a, you know, at the end of the day, everyone was like, wow, that's. great. You know, we went from a bundle size of 200k to a bundle size of 20K, and it was a lot faster and better. So, you know, broadly speaking, I think good engineering cultures, you don't want a culture of, you know, ready, fire aim, but I also think you just need to be really outcomes-oriented. And I think people, if they become, they'd start to treat their code as too precious. It can really impede forward progress. And, you know, I'll just take, like, my understanding, is like a lot of the early self-driving car software was a lot of hand-coded heuristics and
Starting point is 01:39:42 rules and you know a lot of smart people think that eventually will probably be a more monolithic model that encodes many of the same rules you have to throw out a lot of code in that transition but it doesn't mean it's not the right thing to do and so I think in general yeah there might have been some feathers ruffled but at the end of the day everyone's like that's faster and better like let's let's do it you know which is I think the right decision that's awesome and to give you another hypothetical. I want you to share your inner monologue with me as you think through it. So if I told you you have to put 100% of your net worth into a public company today, and you couldn't touch it for at least 20 years, what company would you invest in? And like,
Starting point is 01:40:23 walk me through your thinking. I literally don't know how to answer that question. How would you think about it without giving me an answer? Like, what are the... Yeah, it's a good question. First of all, I'll give you how I'd think about it, but I'm so... Having not been a public company CEO for a couple years, I blissfully don't pay attention as much to the public markets. And in particular right now, it's obviously valuations have gone up a lot. So there's a – but because it's a long-term question, maybe that doesn't matter. I think what I'd be thinking about right now is over the next 20 years, like what are the parts of the economy that will most benefit from this current wave of AI? That's not the only way to invest over 20-year period,
Starting point is 01:41:05 but certainly it's a domain that I understand. And in particular, you know, I mentioned that talk I heard a snippet of from Tyler Cohen, which is like it will probably, AI will probably benefit different parts of the economy disproportionately. There will be some parts of the economy that can essentially where intelligence is a limiting factor to its growth and where you can absorb
Starting point is 01:41:29 almost arbitrary levels of intelligence and generate almost arbitrary levels of growth. Obviously, there's limits to all of this just because you change one part of the economy. It impacts other parts of the economy, and that was what Tyler's point was in his talk. But I would probably think about that because I think that over a 20-year period,
Starting point is 01:41:50 there are certain parts of society that won't be able to change extremely rapidly, but there will be some parts that probably will, and it'll probably be domains where intelligence is the scarce resource right now. And then I would probably try to find companies that will disproportionately benefit from it. And I assume this is why, like,
Starting point is 01:42:09 NVIDIA's stock is so high right now because if you want to sort of get downstream, you know, NVIDIA will probably benefit from all of the investments in AI. I'm not sure I would do that over a 20-year period, just assuming that the infrastructure will shift. So I don't have an intelligent answer, but that's the way I would think about it
Starting point is 01:42:25 if we were doing that exercise. I love that. Where do you think? like what's your intuition say about what areas of the economy are limited by intelligence and not just economy, I mean, perhaps politicians might be limited by this and aid and benefit from, in which case countries could benefit enormously from AI and unlock growth and potential in their economy. But I think maybe just to scope the question, like what areas of the economy do you think are limited by intelligence or workers, like smart workers, in which
Starting point is 01:42:59 case like that's another limit of intelligence yeah i mean uh two that are i think probably going to benefit a lot are technology and finance um you know where you're you know if you can make better financial decisions than competitors you'll generate outsized returns and that's why over the past you know 30 years you know of machine learning um you know uh hedge funds and financial services institutions, everything from fraud prevention to true investment strategies, it's already been an area of domain, domain of investment. Software, similar as we talked about, I think that at some point, we will be, we will no longer be supply constrained in software, but we're not anywhere close to it right now. And you're taking something that has always been
Starting point is 01:43:49 the scarce resource, which is software engineers, and you're making it not scarce. And I think as a consequence. You just think of like how much can that industry grow? We don't know. But we've been so constrained on software engineering as a resource. Who knows over the next 20 years, but we'll find out where the limits are. But to me, intellectually, there's just a ton of growth there. And then broadly, I think areas of like processing information are areas that will really benefit quite a bit here. And so that, and I think the thing that I would think about over 20 period is like second and third order effects, which is why I don't have an intelligent answer. And if you're asking me to put all my money in something, I would think about it for a
Starting point is 01:44:31 while, probably use O1 Pro a little bit to help me. But, you know, because you can end up generated a bunch of growth in the short term, but then, you know, if everyone does it, it commoditizes the whole industry, you know, type of thing. So, you know, there used to be, you know, before the introduction of the freezer, ice was like a really expensive thing and now it's free, you know. And so I think it is really important to actually think through those if you're talking in a time frame of like 20 years. And that's why having not thought about this question ahead of time, you could be quite simplistic elsewhere.
Starting point is 01:45:04 But I would say software and finance are areas that I think stand to reason should benefit quite a bit. I love that response. How do you balance having a young family with also running a startup again? I work a lot. I don't. I really care. and love care about and love working.
Starting point is 01:45:26 So one thing is that I, well, there's always tradeoffs in life. If I didn't love working, I wouldn't do it as much as I do. But I just love to create things and love to have an impact. And so I like jump out of bed in the morning and work out, go to work. And then it's been telling my family broadly, probably being honest. first I'm not perfect but second I don't have a ton of hobbies you know I basically work and spend time with my family
Starting point is 01:45:57 the first time we talked to you saw the couple guitars in my background I haven't picked one of those up in a while I mean I literally pick it up occasionally but I you know do not devote any time into it and I don't regret that either like I am so passionate about what we're building at Sira I'm so passionate about opening I'm so
Starting point is 01:46:17 love my family so much I don't really have any regrets about it but I basically just life is all about where do you spend your time and mine is at work and with family and so that's how I do it I don't know if I'm particularly balanced but I don't strive to be either I really take a lot of pride and I love I love to work having sold the companies you started twice how does that influence what you think of Sierra like are you thinking like oh I'm building this in order to sell it or do you think differently like this is my life's work I'm building this with that's not going to happen? I absolutely
Starting point is 01:46:51 intend Sierra to be an enduring company and an independent company. But to be honest, every entrepreneur with every company starts that way. And so I'm really grateful for both Facebook and Salesforce for having acquired my previous companies and hopefully
Starting point is 01:47:09 I had an impact about those companies. But you don't start off, well, at least I never started off saying, hey, I want to make a company to sell it. But I actually actually think with Sierra we have just a ton of traction in the marketplace. I really do think Sierra is a leader in helping consumer brands build customer facing AI agents. And I'm really proud of that. So I really see a path to that. And I joke with Clay, I want to be an old man
Starting point is 01:47:35 sitting on his porch, you know, complaining how the next generation of leaders at Sierra don't listen to us anymore. You know, I want this to be something that not only is enduring, but outlives me. And I think just actually, I don't think we've ever talked about this, but it was, really interesting moment for me when Google went from its one building in Mountain View to its first corporate campus. We moved into the Silicon Graphics campus, which was right over near Shoreline Boulevard in Mountain View. And
Starting point is 01:48:04 SGI had been a really successful company enough to build a campus. And when we, it was actually quite awkward, we moved into like half the campus. They were still in half and they're like, we're this up-and-coming company, they're declining. And then when Facebook, when we moved, out of the second building. We were in Palo Alto, a slightly larger building.
Starting point is 01:48:24 I think we leased it from HP. But when we finally got a campus, it was from Sun Microsystems who had gone through an Oracle acquisition and had been sort of on the decline. And it was interesting to me because both SGI and Sun had been started and grown
Starting point is 01:48:40 to prominence in my lifetime. Obviously, it was maybe like a little younger, obviously, but in my lifetime, enough to build a whole corporate campus and then declined fast enough to sell that corporate campus to a new software company. And for me, it was just so interesting to have done that twice to move into like a, you know, a used campus for the previous owners.
Starting point is 01:49:04 It was a very stark reminder that technology companies aren't entitled to their future success. And I think we'll see this actually now with AI. AI, I think, will change the landscape of software to be tools of productivity. to agents that actually accomplish tasks. And I think it will help some companies for whom that's a amplifies their existing value proposition and it will really hurt others where it will essentially the seat-based kind of model of legacy software will wane very quickly and then really harm them. And so when I think about what it means to build a company that's enduring, that is a really,
Starting point is 01:49:46 really tall task in my mind right now because it means not only making something that's financially enduring over the next 10 years, but setting up a culture where a company can actually evolve to meet the changing demands of society and technology when it's changing at a pace that is like unprecedented history. So I think it's one of the most fun business challenges of all time. And I think it has as much to do with culture as it has to do with technology because every line of code in Sierra today will be completely different, you know, probably five years from now, let alone 30 years from now. And I think that's really exciting. So when I think about it, I just get so much energy because it's incredibly hard and it's harder now that it's ever been
Starting point is 01:50:33 to do something that lasts beyond you. But that I think is the ultimate measure of a company. You mentioned AI agents. How would you define that? What's an agent? I'll define a more broadly, and then I'll tell you how we think about it, Sierra, which is a more narrow view of it. The word agent comes from agency, and I think it means affording a software, the opportunity to reason and make decisions autonomously. And I think that's really all it means to me, and I think there's lots of different applications of it. The three categories that I think are meaningful, and I'll end with the Sierra one, just so I
Starting point is 01:51:12 can talk about it a little more, but one is personal agents. So I do think that most people will have probably one, but maybe a couple AI agents that they use on a daily basis that are essentially amplifying themselves as an individual. You can do the rote things like help you triage your email to help you schedule a vacation. You know, you're flying back to Edmonton and help you arrange your travel. Two more complex things. Like, you know, I'm going to go ask my boss for promotion, like help me role play. And, you know, I'm setting up my resume for this job, help me do that to, I'm applying
Starting point is 01:51:53 for a new job, help me find companies I haven't thought of that I should be applying to. And I think these agents will be really powerful. I think it might be a really hard product to build because when you think about all the different services and people you interact with every day, it's kind of. of everything. So it's not, it has to generalize a lot to be useful to you. And because of the personal privacy and things like that, it has to work really well for you to trust it. So I think it's going to take a while to go, I think it'll be a lot of demos. I think it'll take a while to be robust. The second category of agent is, I would say, really filling in a persona
Starting point is 01:52:30 within a company. So a coding agent, a paralegal agent, a analyst agent. I think these already exist. I mentioned Cursor. There's a company called Harvey that makes a legal agent. I'm sure there's a bunch in the analyst space. These do a job, and they're more narrow, but they're really commercially valuable because most companies hire people or consultants that do those things already, like analyze the contracts of your supply chain, right?
Starting point is 01:53:01 That's kind of a rote kind of law, but it's really important, and AI can do it really well. So I think that's why this is the area of the economy that I think is really exciting. And I'm really excited about all the startups in this space because you're essentially taking what used to be a combination of people and software and really making something that solves a problem. And by narrowing the domain of autonomy, you can have more robust guardrails and even with current models actually achieve something that's effective enough to be commercially viable today. and by the way it changes the total addressable market of these models too like I don't know what the total addressable market of legal software was three years ago but it couldn't have been that big I couldn't tell you like a legal software company I probably should I just can't think of one but if you think about the money we spend on
Starting point is 01:53:55 lawyers that's a lot and so you end up where you're broadening the addressable market quite a lot the domain we're in I think is somewhat special which is a company's branded customer-facing agent. And the reason why I think it's, one could argue, we're sort of helping with customer service, which is a persona, a role. But I do think it's broader than that because if you think about a website,
Starting point is 01:54:24 you know, like your insurance company's website, try to list all the things you can do on it. You can look up the stock quote. You can look up the management team. You can compare their insurance company to all their competitors. You can file a claim. You can, you know, buy, you can bundle your home and auto.
Starting point is 01:54:43 You can add a member of your family to your premium. There's a million things you can do on it. Essentially, over the past 30 years, websites, a company's website, singular, has come to be the universe of everything that you can do with that company. I like to think it was like the digital instantiation of the company. And that's what we're helping our customers do at Sirius, help them build a conversational AI that does all of that. So, you know, most of our customers start with customer service,
Starting point is 01:55:13 and it's a great application because no one likes to wait on hold and having something that has perfect access to information is multilingual and empathetic is just amazing. But, you know, when you put a conversational AI as your digital front door, people will say anything they want to it. And we're now doing product discovery, consider purchases, going back to insurance example. Hey, you know, I've got a 15-year-old daughter. I really am concerned about the cost of her premium until she grows up.
Starting point is 01:55:44 Tell me which plan I should be on. Tell me why you'll be better than your competitors. That's a really complex interaction, right? That's not something that can you make a web page that does that? No, but that's a great conversation. And so we really aspire that when you encounter. a branded agent in the wild, we want Sierra to be the platform that powers it. And it's super important because there was a case, at least in Canada, where an AI agent for Air Canada
Starting point is 01:56:10 hallucinated a bereavement policy. Right, but they were found liable to hold themselves to what the agent said. Yeah, I mean, it turns out... And it was an AI agent. There was no human involved in the whole thing. Well, look, it's one thing if chat GPT hallucinate something about your brand. It's another if your AI agent, who they say something about your brand. So the bar just gets higher. So the robustness of these agents, the guardrails, everything is more important when it's yours and it has your brand on it. And so it's harder, but I also, I'm just so excited for it because this is a little
Starting point is 01:56:47 overly intellectual, but I really like the framing. If you think about a modern website or mobile app, it's essentially you've created a directory of functionality from which you can choose. But the main person with agency and that is the creator of the website. Like what are the universe of options that you can do? When you have an AI agent represent your brand, the agency goes to the customer. They can express their problem any way they want in a multifaceted way. And so it means that like your customer experience goes from the enumerated set of functionality you've decided upon your website to whatever your customers ask. And then, you know, you can decide how to fulfill those requests or whether
Starting point is 01:57:31 you want to. Yeah, yeah. But I think it would really change the dynamic to be really empowering to consumers. As you said, I mean, the reason that Air Canada case is the reason we exist. You know, companies, if they try to build this themselves, there is a lot of ways you can shoot yourself in the foot. But in particular, too, your customer experience should not be wedded to one model, let alone even this current generation of models. So with Sierra, you can define your customer experience once in a way that's abstracted from all of the technology. And it can be a chat.
Starting point is 01:58:08 It can be, you can call you on the phone. It can be all of those things. And as new models and new technology comes out, our platform just gets better. But you're not like re-implementing your customer experience. And I think that's really important because we were talking about what's happened over the past two years. Can you imagine if you're a consumer brand like ADT home security and thinking about like how can you maintain your AI agent in the face of all of that, right? It's just not even, it's not tenable. I mean, it's not what you do as ADT. So they've worked with us to build their AI agent. How do you fend off
Starting point is 01:58:39 complacency? Like a lot of these companies and maybe not in tech specifically, but they get big, they get dominant, and then they take their foot off the gas. And that open. opens the door to competitors. And there's like a natural entropy almost to bureaucracy in some of these companies that and the bureaucracy serves the seeds of failure and competition. How do you, how do you fend that off constantly? It is a really challenging thing to do at a company. One of the, there's two things that I've observed that I think manifest as corporate complacency.
Starting point is 01:59:18 One is bureaucracy. And I think the root of bureaucracy is often, when something goes wrong, companies introduce a process to fix it. And over those like sequence of 30 years, the layered sum of all of those processes that were all created for good reason with good intentions, end up being a bureaucratic sort of machine where the reason. the reasons for many of the rules and processes are rarely even remembered by the organization but it creates this sort of natural inertia sometimes that inertia can be good
Starting point is 02:00:02 you know it's like you know if you end up with there's definitely been stories of executives coming in and ready fire aim new strategies that backfire massively but often it can mean in the face of a technology shift or a new competitor you just can't move fast enough to address it the second thing that I think is more subtle is as a company grows in size, often its internal narrative can be stronger than the truth from customers.
Starting point is 02:00:30 I remember one time when this sort of peak of the smartphone wars and I end up visiting a friend on Microsoft's campus. And I got off the plane in Seattle-Tacoma Airport, drove into Redmond, went onto the campus. and all of a sudden everyone I saw was using Windows phones. I assume it must have been a requirement or formal or social. You were definitely uncool if you were using anything else. And from my perspective, at the time, like the war had already been lost. Yeah.
Starting point is 02:01:06 Like it was definitely a two-horse race between Apple and Google on iOS and Android. And I remember sitting in the lobby waiting for my first. friend to get me from the security check-in. And I made a comment, like, it wasn't a confrontation about, but I made a comment to someone who was at Microsoft, I was like, you know, something along the lines, are you required to use Windows phones, how these others? And I just sort of like curious. And then I got a really bold answer, which is like, yeah, we're going to win. Like, we're taking over the smartphone market. And I was like, I didn't say anything because it was like a little socially
Starting point is 02:01:42 awkward of like, no, you're not. Like, you lost like four years ago. But there's a process, there's something that's happening that's preventing you from getting reality. Well, and that's the thing is if you think about it, if anyone, if you've ever worked for like a large company, you know, you're, when you work at a small company, you care about your customers and your competitors and you feel every bump in the road, when you're a, you know, junior vice president of whatever and you're, you know, eight levels below your, you know, CEO. and you have a set of objectives and results. You might be focused as I want to go from junior vice president to senior vice president. That's what success looks like for me. And you end up with this sort of myopic focus on this internal world. In the same way your kids will focus on, you know, the social dynamics of their high school,
Starting point is 02:02:37 not the world outside of it. And it's probably rational, by the way, because like, you know, probably their social life is more determined by those. you know 1,000 kids in their high school and it's like all the things outside but this is that's the life of a person inside of these big places and so you end up where uh you know if you have a very senior head of product who's like are this competitor says they're faster but this next version we're so much better and then everyone says and all of a sudden that's like the windows phone is going to win that's what everyone says and and you truly believe it because everyone you meet
Starting point is 02:03:11 says the same thing and you end up reflecting, you know, customer anecdotes through that lens and you end up with this sort of reality distortion field manifested from the sum of this sort of myopic storytelling that exists within companies. What's interesting about that is like, you know, the ability for a culture to believe in something is actually a great strength of a culture, but it can lead to this as well. And so the combination of bureaucracy and inaccurate stories, storytelling, I think, is the reason why companies sort of die. And it's really remarkable to look at, you know, the BlackBerrys of the world or the TiVos or that, you know, you can really, you know, as the plane is crashing, like, tell the story that you're not. And, and, and then similar, as I said, like, culturally, you can still have, like, the person in the back of that crashing plane being like, when am I going to get promoted to SVP? and you're like, what the, you know. Yeah.
Starting point is 02:04:13 And that's, I mean, this is like, I mean, I've seen it a hundred times. And so I think it really comes down to leadership, you know, and I think that one of the things that most great companies have is they are obsessed with their customers. And I think the free market doesn't lie. And so I think the one of the most important things I think for any like enduring culture, particularly in an industry that changes as rapidly as software is how close are your employees to customers and how much. can customer, like the direct voice of your customers, be a part of your decision-making? And that is something that I think you need to constantly work out because that, you know, person, employee number 30,000, 462, you know, how does he or she actually directly hear from customers?
Starting point is 02:05:01 It's not actually a simple question to answer. Is it direct? Is it filtered? How many filters are there? That's exactly right. And then I think the, other part on leadership is, you know, we talk about bureaucracy is process is there to serve the needs of the business. And often mid-level managers don't get credit for removing process. They often are held accountable for things going wrong. And I think it really takes top-down leadership to, you know, remove bureaucracy. And it is not always comfortable, you know, when companies remove spans of control or all the people impacted, well, it's like antibodies.
Starting point is 02:05:49 And for good reason, it makes sense. Their lives are negatively impacted or whatever it is. But it almost has to come from the top because you need to give air cover. Almost certainly something will go wrong, by the way. I mean, like processes usually exist for a reason. But when they accumulate without end, you end up with bureaucracy. So those are the two things that I always, and you could smell it when you go into a really bureaucratic company, the inaccurate storytelling, the process over outcomes. And it's just, it sort of sucks the energy out of you and you feel it.
Starting point is 02:06:24 That's a great answer. We always end these interviews with the exact same question, which is, what is success for you? Success for me. We talked about how I spend my time with my family at work is, you know, having a happy, healthy family. and being able to work with my co-founder, Clay, for the rest of my life, making Sierra into an enduring company. That would be success for me. Thanks for listening and learning with us.
Starting point is 02:06:53 The Farnham Street blog is where you can learn more about my new book, Clear Thinking, Turning Ordinary Moments into Extraordinary Results. It's a transformative guide that hands you the tools to master your fate, sharpen your decision-making and set yourself up for unparalleled success. Learn more at fs.blog slash clear. Until next time.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.