a16z Podcast - AI Revolution: Top Lessons from OpenAI, Anthropic, CharacterAI, & More

Episode Date: September 25, 2023

The AI Revolution is here. In this episode, you’ll learn what the most important themes that some of the world’s most prominent AI builders – from OpenAI, Anthropic, CharacterAI, Roblox, and mor...e – are paying attention to. You’ll hear about the economics of AI, broad vs specialized models, the importance of UX, and whether we can expect scaling laws to continue.This footage is from an exclusive event, AI Revolution, that a16z ran in San Francisco recently. If you’d like to access all the talks in full, visit a16z.com/airevolution. Topics Covered00:00 - AI Revolution01:42 - The economics of AI06:55 - The third epoch of compute13:52 - Recognizing scaling laws17:42 - Can scaling laws continue?22:34 - Potential bottlenecks25:58 - Personalization vs generality29:43 - The importance of UX31:55 - The future of multi-modality Resources:Catch the all the talks at https://a16z.com/airevolution Stay Updated: Find a16z on Twitter: https://twitter.com/a16zFind a16z on LinkedIn: https://www.linkedin.com/company/a16zSubscribe on your favorite podcast app: https://a16z.simplecast.com/Follow our host: https://twitter.com/stephsmithioPlease note that the content here is for informational purposes only; should NOT be taken as legal, business, tax, or investment advice or be used to evaluate any investment or security; and is not directed at any investors or potential investors in any a16z fund. a16z and its affiliates may maintain investments in the companies discussed. For more details please see a16z.com/disclosures.

Transcript
Discussion (0)
Starting point is 00:00:00 Look, this is the beginning of something amazing because there's no limit. This is right now an inflection point where we're sort of redefining how we interact with digital information. These are the fastest-going open-source projects. These are the fastest-going products and some of the fastest-going companies we've seen in the history of the industry. We, for a long time, really focused on building our own infrastructure. We have hundreds of thousands of servers. He said, well, I think we can get by with like 500. And said, okay, I think we can find 500K somewhere.
Starting point is 00:00:30 And I remember your dead band thing. Dude, I'm talking about $500 million. The internet was the dawn of universally accessible information, and we're now entering the dawn of universally accessible intelligence. The AI revolution is here. But as we collectively try to navigate this game-changing technology, there are still many questions that even the top builders in the world are grappling to answer. And that is why A16 and Z recently
Starting point is 00:01:00 brought together some of the most influential founders from Open AI, Anthropic, Character AI, Roblox, and more, to an exclusive event called AI Revolution in San Francisco recently. And in today's episode, we share the most important themes from this event, starting with the economics of AI, but we also touch on broad versus specialized models, and which ultimately may win, the importance of UX, and also whether we can expect scaling laws to continue. By the way, several founders comment on what they're seeing there, including Noam Shazir, lead author of the preeminent Transformer paper from back in 2017. Now, I won't delay us any longer, other than saying we've got a lot more coverage of the spent coming,
Starting point is 00:01:45 including how AI is disrupting everything from games to design, how two important waves in machine learning and genomics are colliding, and what we can expect from the enterprise. But in the meantime, if you would like to listen to all the talks, full today, you can head on over to A16Z.com slash AI Revolution. As a reminder, the content here is for informational purposes only, should not be taken as legal, business, tax, or investment advice, or be used to evaluate any investment or security and is not directed at any investors or potential investors in any A16C fund.
Starting point is 00:02:21 Please note that A16D and its affiliates may also maintain investments in the companies discussed in this podcast. For more details, including a link to our investments, please see A16C.com slash Disclosures. All right, let's start with Martine Casado, general partner at A16CZ, giving the why now, and also how the economics of the space may finally be coming together. I will give you the punchline up front. The punchline is if you've ever wanted to start a startup or join a startup,
Starting point is 00:02:55 now is a great time to do it. But how early are we in the trajectory of the same? technology. For example, the microchip was invented in the late 50s, but it wasn't until the turn of the century when Steve Jobs famously put a thousand songs in your pocket. So just how much opportunity is still in the table. Okay, so what has the narrative been for AI over the last 50 years? The narrative is this episodic thing with summers and winters and all of these false promises. I remember when I joined PhD in 2003, for my cohort that joined, I would say 50% of the people were doing AI. This was like when Bayesian stuff was super popular. And then within three years,
Starting point is 00:03:38 everybody's like, AI is dead, right? And so it's kind of been with this love, hate relationship with it for a very long time. But if you look at all of the graphs, we've made tremendous amount of progress for the last 70 years. And along the way, we've solved a lot of very real problems, right? Like way back in the 60s, we're doing expert systems, which are still used for diagnosis, right? Like we're very good at beating Russians at chess. You know, we're doing self-driving cars, we're doing vision. There's just a lot of stuff that we've solved. And so much so it's become a cliche that every time we solve a problem, we're like, oh, well, that wasn't really AI, right? So we just keep moving the goalposts. So we've had steady progress. We've solved real
Starting point is 00:04:12 problems. And not only that, it's been a while now that we've been better at humans for some very important things, for example, like perception or handwriting detection. It's been about 10 years since we've been better than humans at entity identification. And not only, that, we've actually gotten very good at monetizing this, particularly for large companies, right? And so, as we all know, there's been a ton of market cap that's been added to companies like Meta and Google and Netflix by using AI. So I think the question we should all ask ourselves is why hasn't this resulted in an actual platform shift? And by platform shift, I mean, why has the value accrued to the incumbents? And why hasn't there been a new set of kind of AI native companies
Starting point is 00:04:53 that have come up and displaced them, which we've seen in many other areas, right? We saw that in mobile, obviously we saw it with the microchip, et cetera. But I'm going to argue is that the capabilities have all been there, but the economics just haven't for startups. So if you step back and you look at the standalone case for AI economics, not like what a big company can extract from it, but a startup, it's actually not been great. I mean, to begin with, a lot of the sexier use cases are pretty niche markets, right?
Starting point is 00:05:19 Like, you know, it's great to beat Russians at chess. like maybe it's a useful tool that you can apply to solving bigger problems, but that not itself is a market. I actually think the second point is the most important point and is pretty subtle. Many of the traditional use cases of AI require correctness in the tail of the solution space. And that's a very hard thing for a startup to do for a couple of reasons. One of the reasons is if you have to be correct and you've got a very long and fat tail, either you do all of the work technically or you hire people.
Starting point is 00:05:50 So often we hire people, right? and for startups to start hiring people to provide solutions as a variable cost. And the second one is because the tails of these solutions tend to be so long, think something like self-driving, where there's so many exceptions that could possibly happen, the amount of investment to stay ahead increases and the value decreases, right? You have this perverse economy of scale. So we've actually done studies on this, and it turns out many companies that try to do this as startups end up with non-softrow-like margins.
Starting point is 00:06:14 They're lower margins, and they're just much harder to scale. Of course, with robotics comes to curse of hardware. classically a very difficult thing for startups to do. And if you really think, like, what is the competition of most use cases of AI? It tends to be the human, and traditionally it's stuff like the human brain is really good at, like perception, right? Like the brains that we have have evolved over 100 million years to do things like whatever, pick berries and evade lions, or whatever it is, and it's incredibly efficient at doing that. So this leads to something that most investors know, which we call the dreaded AI mediocrity spiral. And what is it? It's very
Starting point is 00:06:47 simple, which is, let's say a founder comes in and they want to do an AI company and they're going to use AI to automate a bunch of stuff. Of course, correctness is really important and they wanted to look at it first so they hire people to do it instead of the AI. Then they come to us. We invest in them and I joined the board. Then I say, listen, this is great. You need to grow. And they're like, oh, man, we need to grow. This AI's hard. Like the tail's very long. I'm going to hire more people. And now you're on this treadmill of continuing hiring people. And this is one of the reasons why so many startups that have tried to do this just haven't had kind of this breakaway economics, and the value accrues to large companies that can actually seek these perverse economies of
Starting point is 00:07:23 scale. But, you know, market transformations aren't created when economics get 10 times better. They get created when they're 10,000 times better. So what is the learning from the last, say, 70 years? It's not that the technology doesn't work. It's not that we can't solve the problems. It's not even that we can't monetize it. Big companies are great at monetizing it. It's that it's very, very hard for startups to break away. And if startups can't break away, you don't get a transformation. But what about the current wave, where the everyday consumer can prompt LLMs with natural language and have an output of variety of things, from conversations to images to even 3D models? So this wave is very, very different, and we're already
Starting point is 00:08:05 seeing productive viable businesses, right? Like I like to call them the 3C's. There's creativity, like any component of a video game, you can automatically. generate. There's companionship, which kind of more of emotional connections. And then there's a class that we call co-pilot, which will help you with tasks. These are already emerging as independent classes. So remember the properties of AI previously that made it difficult to build a startup company. So none of these really apply to this current model. The first one, obviously, these are large markets that this is being applying to. It's like arguably all of white collar work, even just like video games and movies is like $300 billion market. These are massive,
Starting point is 00:08:44 massive markets. The second one, again, I think is, you know, the most important point and maybe the most subtle. In this domain, correctness isn't as much of an issue for two reasons. One of them is when you're talking about creativity, the first C, there is no formal notion of correctness, really. I mean, like, what does it mean to be incorrect for like a fiction story or a video game? I mean, for sure, like, you want to make sure, like, they have all their fingers, but even then, do you really? In sci-fi? And so we have absolutely adapted to use cases where, you know, correctness is not a huge issue. The second one is a little more subtle, I just think it's so important,
Starting point is 00:09:16 which is the behavior that's developed around these things is iterative. And so that human-in-the-loop that used to be in a central company is now the user. So it's not a variable cost to the business anymore. The human-in-the-loop has moved out. And as a result, you can do things where correctness is important. Like, for example, developing code because it's iterative. so like you're constantly getting feedback and correction from the user. And I want to talk to about this brain portion
Starting point is 00:09:39 because I think it's so interesting. I'm not a neuroscientist, but for these types of tasks, the silicon stack is way better than the carbon stack, right? So if you think about it, like traditional AI, a lot of it is doing stuff like the 100 million-year-old brain is doing, right? The one that's been fleeing predators or like picking strawberries or whatever it is.
Starting point is 00:09:56 And that's very, very hard to compete with. Like remember, if you have the CPU-GPU set up a self-driving car, Some of these kits are like 1.3 kilowatts where the human brain is 15 watts. So economically, that's very tough to compete with. The new Gen I Wave, it's kind of competing with the creative language center of the brain. It's like 50K years old. Like it's much less evolved. And it turns out it's incredibly competitive,
Starting point is 00:10:18 so much so that you actually have the economic inflection we look for for a market transformation. So let's just go ahead and break down the numbers very quickly. So let's say that I, Martine, wanted to create an image of myself as a Pixar character. So if I'm using one of these image models, the inference cost, let's call it, you know, a tenth of a penny. It's probably less than that, actually. Let's say it takes one second. If you compare that to hiring a graphic artist, let's say that it was a hundred bucks in an hour. I've actually hired graphic artists to do things like this.
Starting point is 00:10:44 It tends to be a lot more money than that, but let's conservatively say that. You've got four to five orders of magnitude difference in cost and time. So these are the type of inflexions you look for certainly as an economist when you're like there's going to be actually a massive market dislocation. I'll give you another example from Instabase. So let's assume that you have a legal brief. It's in a PDF. You throw it into this kind of unstructured document, LLM, and then you ask questions for that legal brief.
Starting point is 00:11:10 Again, the inference costs, say a tenth of a penny. Maybe it's a little more, maybe it's a little less. Time to complete, maybe one second, maybe a little more, maybe a little less. But as someone who has actually spent a lot of money on lawyers hours, I want to point out a couple of things. The first one is it takes more than one hour to iterate on this for sure. And the second one is they're not always correct. In fact, built in, for any interaction I have,
Starting point is 00:11:28 with a lawyer is cross-checking their work and double-checking the work. So again, we have four to five orders of magnitude difference in cost and time. And if you want to have like an example of like how the extremely nutty that this can get, like I see no reason why you can't generate an entire game. There are companies today working on it, the 3D models, the characters, the voices, the music, the stories, et cetera. Like there are companies that are doing all of these things. And if you compare the costs of like hundreds of millions of dollars and ears
Starting point is 00:11:54 versus, you know, a few dollars of inferences, now we have like current, like internet and microchip level asymmetries and economics. Now, listen, I'm not going to say this happens soon. We're not there yet. What I'm saying is this is the path that we're on. And these types of paths are what you look for with big transformations. So it's little wonder why we're seeing so much takeoff the way that we have. And these are the fastest-going open-source projects.
Starting point is 00:12:19 These are the fastest-going products and some of the fastest-going companies we've seen in the history of the industry. And it's because, again, it's less the capabilities, and it's much more that the economics work. So, listen, this may sound hyperbolic, but I really think that we could be entering a third epoch of compute. I think that the first epoch, of course, is the microchip.
Starting point is 00:12:37 Before the advent of the computer, you actually had people calculating longer than tables by hands. Like, that's where the word comes from. They were computers. They would compute. Then we created ENIAC, along with other machines, but let's look at ENIAC. So ENIAC was 5,000 times faster than a human being doing it. There's your three to four orders of magnitude,
Starting point is 00:12:53 and that kind of ushered in the compute revolution, right? And this gave us a number of companies that were either totally transformed like IBM or totally net new. So the microship brought the marginal cost of compute to zero. The internet brought the marginal cost of distribution to zero. So listen, in the 90s, when I wanted to get a new video game, I would go to a store and buy a box.
Starting point is 00:13:13 And again, I don't have the math up here, but if you actually calculate the price per bit relative to DSL in the late 90s, is about four or five orders of magnitude again, relative to actually shipping it. So I think it's a pretty good analog where you say these large models actually bring the marginal cost of creation. There was some very fuzzy, vague notion of what creation means. But for sure, we could talk about it of like content, conversation, whatever it is.
Starting point is 00:13:35 And like the previous epochs, when those epochs happened, you had no idea what new companies were going to be created. Nobody predicted Amazon, nobody predicted Yahoo. Like, I remember when this happened. So I think, listen, I think we should all get ready for a new wave of iconic companies. I don't think we know what they're going to look like. But forget the capabilities. Economics are just too compelling. We'll hear more from our teen at the end of this episode.
Starting point is 00:13:59 But, speaking of economics and the scale of top models today, here is our new general partner, Anjini Midha, reminiscing about an early call he had with Dario Amadeh, co-founder of Anthropic, who you'll also hear from shortly. I'm going to take you all back in time to about three years ago. You and Tom gave me a call, one of your co-founders, and said, hey, we think we're going to go start Anthropic. And I asked you, great, okay, like, what do you think we need to get going?
Starting point is 00:14:25 And he said, well, I think we can get by with, like, 500. I said, okay, I think we can find 500 days somewhere. And I remember you deadband saying, dude, I'm talking 500 million dollars. And that's when I realized that things were going to be a little bit different. Dario was one of the first employees at Open AI and spent five years there before co-founding Anthropic. And the last year of AI has absolutely captured the masses, but people like Dario, were early in recognizing just how far these technologies could scale. What was it at that moment when the team at OpenA had started publishing
Starting point is 00:15:00 your first experiments on scaling laws that gave you so much confidence that this was going to hold when everybody else just thought that was crazy talk? Yeah, so for me, the moment was actually GPT2 in 2019, where there were two different perspectives on it, right? When we put out GPT2, you know, some of the moment. the stuff that was considered most impressive at the time was, oh, my God, you give this five examples, just offer it straight into the language model, five examples of English to French translation, and then you put a six sentence in English, and it actually translates it into French,
Starting point is 00:15:36 like, oh, my God, it actually understands the pattern. That was crazy to us, even though the translation was terrible, it was almost worse than if you were to just take a dictionary and substitute word for word. But, you know, our view was that, look, this is the beginning of something amazing because there's no limit and you can continue to scale it up and there's no reason why the patterns we've seen before won't continue to hold. The objective of predicting the next word is so rich
Starting point is 00:16:01 and there's so much you can push against that it just absolutely has to work. And then some people looked at it and they're like, you made a bot that translates really badly. It was just, I think, two very different perspectives on the same thing and we just like really, really believed in the first perspective. What happened then was you saw a reason to continue
Starting point is 00:16:20 down that line of inquiry, which result in GPD3. And what would you think was the most dramatic difference between GPD3 and the previous efforts? Yeah, I mean, I think it was much larger and scaled up to a substantial extent. I think the thing that really surprised me was the Python programming, where the conventional wisdom was that these models couldn't reason at all. And when I saw the Python programming, even though it was very simple stuff, even though a lot of it was stuff you could memorize, you know, you could put it in.
Starting point is 00:16:50 kind of new situations, come up with something that isn't going to be anywhere in GitHub, and it was just showing the beginning of being able to do it. And so I felt that that ultimately meant that we could keep scaling the models, and they would get very good at reasoning. What was the moment at which you realized, well, okay, we think this is actually going to generalize much broader than we expect? What were some of the signals there that gave you that conviction? I think one of the signals was that we hadn't actually done any work.
Starting point is 00:17:16 We had just scraped the web, and there was enough. Python data in the web to get these good results. When we looked through it, it was like maybe 0.1% to 1% of the data that we scraped was Python data. So the conclusion was, well, if it does so well with so little of our data and so little effort to curate it on our part, it must be that we can enormously amplify this. And so that just made me think, well, okay, we're getting more compute. We can scale up the models more, and we can greatly increase the amount of data.
Starting point is 00:17:46 So we have so many ways that we can amplify this. And so, of course, it's going to work. It's just a matter of time. Another person optimistic about scaling laws at the time was Noam Shazir. Noam was one of the researchers and lead author behind the transformative 2017 transformer paper and has since co-founded Character AI. I knew that, you know, you can make this technology better in a lot of ways. We can improve it with model architecture and distributed algorithms and quantization and all of these things.
Starting point is 00:18:16 So I was working on that, but then struck me, hey, The biggest thing is just scale. Can you throw like a billion dollars or a trillion dollars at this thing? What would happen if we did massively scale compute? Well, many companies chose to find out, and we, the consumer, are the beneficiaries of that. But can this realistically continue? Can the industry just continue to throw more computer at the problem and get better solutions? Or will a more fundamental unlock be required?
Starting point is 00:18:45 This theme was top of mind for many at the event. And here is OpenAI's co-founder and CTO, Mira Mirati, tackling that question head on. Do you think the scaling laws are going to hold and we're going to continue to see advancements, or do you think we're hitting diminishing returns? So there isn't any evidence that we will not get much better, much more capable models as we continue to scale them across the axis of data and compute. Whether that takes you all the way to AGI or not, that's a different question. There are probably some other breakthroughs and advancements needed along the way.
Starting point is 00:19:23 But I think there's still a long way to go in the scaling laws and to really gather a lot of benefits from these larger models. We'll hear more from Mira and touch on AGI in part two. But first, here's Nome again, in conversation with A66 and Z general partner, Sarah Wayne. on just how much compute we expect to soon be available, but also how much innovation is on deck, even if there aren't additional fundamental breakthroughs. And for those listening on audio,
Starting point is 00:19:52 yes, Nome fully did this computation in his head. I see this stuff like massively scaling up. It's just like not that expensive. I think I saw an article yesterday, like Nvidia is going to build like another one and a half million H100 next year. So like that's 2 million H100. So that's 2 times 10 to the 6 times they can do about 10 to the 15th operations per second. So 2 times 10 to the 21 divide by like 8 times 10 to the 9 people on Earth.
Starting point is 00:20:22 So that's roughly a quarter of a trillion operations per second per person, which means that could be processing on the order of like one word per second on like a 100 billion parameter model for everyone on Earth. But like really it's not going to be everyone on Earth because like some people are blocked in China and some people are sleeping. It's not that expense. You know, like, this thing is massively scalable if you do it right. And, you know, we're working on that.
Starting point is 00:20:49 You said this once, that the internet was the dawn of universally accessible information. And we're now entering the dawn of universally accessible intelligence. Maybe building off your last answer, what did you mean by that? Do you think we're there yet? Yeah, I mean, I think it's like we're really a Wright Brothers first airplane kind of moment, right? Like, we've got something that works and is useful for now. some large number of use cases and looks like it's scaling very, very well. And without any breakthroughs, it's going to get massively better as everyone just kind of scales up to use it.
Starting point is 00:21:21 And there will be more breakthroughs because now, you know, like all the scientists in the world are like working on making this stuff better. It's great that like all this stuff is accessible, open source. Like, you know, we're going to see like a huge amount of innovation and what's possible in the largest companies now can be possible in, you know, in somebody. academic lab or garage in a few years. And then as the technology gets better, there's just going to be all kinds of great use cases that emerge and pushing technology forward, pushing science, pushing the ability to, you know, help people in various ways. I love to get to the point where you can just ask it how to cure cancer or something. I mean, it seems a few years away for now. Do you think we need another fundamental breakthrough like the transformer technology to
Starting point is 00:22:07 get there? Or do you think we actually have everything that we need? I mean, it's impossible to predict the future, but I don't think anyone's seen like these scaling laws, you know, stop. I think as far as anybody has experimented, stuff just keeps getting smarter. So we'll be able to unlock lots and lots of new stuff. I don't know if there's an end to it, but at least everybody in the world should be able to talk to something like really brilliant and have incredible tools all the time. And I can't imagine that that will not be able to build on itself. At the core, the computation isn't that expensive. Like, operations cost like $10 to the negative $18 these days.
Starting point is 00:22:46 And, like, you know, if you can do this stuff efficiently, even talking to the biggest models ever trained, the cost of that should be, like, way, way lower than the value of your time or most anybody's time. And really, we should, you know, there's the capacity there to scale these things up by orders of magnitude. As the industry does pursue scale, here's Dario's take on what bottlenecks may be along the way. With the next 24-36 months, what do you think the biggest bottlenecks are in demonstrating that the scaling laws continue holding?
Starting point is 00:23:18 Yeah, so I think there's three elements. There's data. There's compute and there's algorithmic improvements. So I think we are on track, even if there were no algorithmic improvements from here, even if we just scaled up what we had so far, I think the scaling laws are going to continue, and I think that's going to lead to amazing improvements. I think the biggest factor is simply that more money is being poured into it. The most expensive models made today cost about $100 million, say plus or minus, a factor of two. I think that next year we're probably going to see from multiple players models on the order of $1 billion. And in 2025, we're going to see models on the order of several billion, I don't know, perhaps even $10 billion.
Starting point is 00:24:02 And so I think that that factor of 100, plus the compute inherently getting faster with the H-100s, that's been a particularly big jump because of the move to lower precision. So you put all those things together. And if the scaling laws continue, there's going to be a huge increase in capabilities. But if compute does increase, how might this impact the size of models and ultimately the cost of inference for consumers? Inference will not get that much more expensive. The basic logic of the scaling laws is that if you increase compute by a factor of n, you could increase data by a factor of square root of n,
Starting point is 00:24:42 and size the model by a factor of square root of n. So that square root basically means that the model itself does not get that much bigger, and the hardware is getting faster while you're doing it. So I think these things are going to continue to be servable for the next three or four years. If there's no architectural innovation, they'll get a little bit more expensive. I think if there's architectural innovation, which I expect there to be, they'll get somewhat cheaper. Increased model size and performance should unlock fundamentally new applications, which we'll explore further in part two.
Starting point is 00:25:14 But first, entering the conversation is David Bazuki, co-founder and CEO of Roblox, commenting on the value of owning your infrastructure and the impact of that on inference cost, especially in a 3D world constantly reinventing itself. An even further extension with, it takes probably a lot of compute horsepower, which is completely personalized degeneration in real time, backed by massive inference stuff. So you could imagine, okay, I'm making the Super Dungeons and Dragons thing,
Starting point is 00:25:48 but as it watches you play and maybe we know your history, you'll be playing a 3D experience that's no one's ever seen before. One of the good things we've done is we, for a long time, I'm really focused on building our own infrastructure. We have hundreds of thousands of servers, many, many edge data centers, terabits of connectivity that we've traditionally used for 3D simulation, that the more we can run inference jobs on these,
Starting point is 00:26:13 we can run super high volume inference at high quality, at low cost, and make this just freely available so the creators don't worry about it. Whether we can continue scaling is one thing. But another topic on the minds of, many builders is whether they can compete with the largest models. Will bigger models always win or will specialization trump generality? Martin and Mira discuss. It reminds me very much of the silicon industry. So I remember in the 90s, when you buy a computer, there are all these weird co-processors. There's like, here's string matching. Here's a floating point. Here's crypto. And like all of
Starting point is 00:26:51 them got consumed into basically the CPU. It just turns out generality was very powerful. And that created a certain type of economy one where like you had you know intel and amd and like you know it all went in there and of course creates a lot of money to build these chips and so like you can imagine two futures there's one future where like you know generality is so powerful that over time the large models basically consume all functionality and then there's another future where there's going to be a whole bunch of models and like things fragment and you know there are different points of the design space. Do you have a sense of like, is it open AI and nobody or is it everybody? It kind of depends what you're trying to do. So obviously the trajectory is one where these
Starting point is 00:27:34 AI systems will be doing more and more of the work that we're doing and they'll be able to operate autonomously, but we will need to provide direction and guidance and oversee. But I don't want to do a lot of the repetitive work that I have to do every day. I want to focus on other things. But in terms of like how this works out with a platform, we make a lot of models available through our API from the very small models to our frontier models. And people don't always need to use the most powerful, the most capable models. Sometimes they just need a model that actually fits for their specific use case. And that's far more economical. So I think there's going to be a range. And there's a lot of focus right now on building more models. But, you know,
Starting point is 00:28:20 building good products on top of these models is incredibly difficult. Plus, each industry may have unique requirements. Here's David commenting on how a suite of models will likely be required in order to power the class of games of the 21st century. In any company, like a Roblox, there's probably 20 or 30 end-user vertical applications that are probably very bespoke. Natural language filtering very different than generative 3D. And at the end user point, we want all of those running.
Starting point is 00:28:52 We want to use all of the data in an opt-in fashion to help make these better, tune these better. But as we go down, down, down, there's probably a natural two or three clustering of general, bigger, fatter-type models in a company like ours. There's definitely one around safety, civility, natural language processing, natural language translation.
Starting point is 00:29:14 Generally, one more multimodal thing around 3D creation, say some combination of text, image, whatever, generate a great avatar. And then there's probably a third area, which gets into the virtual human area, which is how would we take the 5 billion hours of human opted-in data, what we're saying, how we're moving, where we go together,
Starting point is 00:29:38 how we work in a 3D environment, and could we use that to maybe inform a better 3D simulation of a human? So I would say, yes, look at large models in those three areas. And I think the market, as we see it, there's going to be these super big God model, massive LLM-type companies.
Starting point is 00:29:58 I think we are probably a layer below that. We're very fine-tuned for the disciplines we want. And it's worth noting that the back-end model is only one part of the product. Here is Mira with a reminder to builders about the importance of UX. Actually, you can see sort of the contrast between making this models available through an API and making the technology available through chat GPT.
Starting point is 00:30:26 It's fundamentally the same technology, maybe with a small difference with reinforcement learning with human feedback for chat GPT. But it's fundamentally the same technology and the reaction and the ability to grab people's imagination and to get them to just use the technology every day is totally different. Here is David Bazuki again, in conversation with A16Z general partner John Lai on what UX may be required, especially given the sheer number of games and experiences that we expect to be enabled with AI. Do you think you'll need to have a new user interface or discovery mechanism?
Starting point is 00:31:07 I think the user interface, there's a lot of opportunity in addition to thinking of this just as content and thinking of this as your real-time social graph. It's fascinating because I think one of the examples of AI being used by big companies is Netflix and then I think TikTok as well if they're sort of very personalized feeds and recommendations. And you could maybe imagine a feature where a user that onboards in the Roblox doesn't actually see a library or a catalog of games, but it's just presented with like a feed.
Starting point is 00:31:35 And it's almost like you're just going from one game. This is really right. I think we're, we are constantly. lead testing, you know, the new user experience. Should that be 2D? Should that be 3D? What's the waiting between creating your digital identity versus discovery? What's the waiting between connecting with your friends and optimizing all that? And we may find that it has to be personalized. Having text or voice prompt is just something that's naturally part of any experience wherever you go. Just like in a traditional avatar editor rather than sliders and radio buttons, that will move
Starting point is 00:32:11 to, I think, a more interactive type of text prompt thing. As we think about UX and the increasing capabilities of these models, how might they let us further integrate with the world around us and connect to more data streams? Here is Mira, David, and Nome, exploring the world of multi-modality. Today, obviously, have this great representation of the world in text, and we're adding other modalities, like images and video and various other things. So these models can get a more comprehensive sense of the world around us,
Starting point is 00:32:46 similar to how we understand and observe the world. The world is not just in text, it's also in images. Yeah, I think there's a lot of interesting stuff going on in various ecosystems around this co-pilot notion. There is one co-pilot where we're all wearing, like, our little earbud all day long, and that co-pilot is talking to us, So that's maybe more consumer real-time copilot. There's obviously many companies trying to build the co-pilot
Starting point is 00:33:14 that you hook up to your email, your text, your Slack, your web browser, and whatever, and it starts acting for you. I'm really interested in the notion that co-pilots will talk to other co-pilots using like natural English, I think will be the universal interface of co-pilots. And you could imagine NPCs being created by prompts, you know, hey, I'm building the historical constitutional thing. I want George Washington there, but I want George Washington to act at the highest level of civility
Starting point is 00:33:47 and utter new users through the experience, tell them a little about constitutional history, go away when they're done. I actually do think you will see those kind of assistance. I mean, also multimodal. Maybe you want to hear a voice and see a face, and then also just able to interact with multiple, people like, yeah, do you want a virtual person, like, in there with, you know, say, with all
Starting point is 00:34:12 your friends, or do you want the experience? It's like you got elected president, you get the earpiece, and you get like the whole cabinet of friends or advisors, or it's like, you know, like you walk into cheers and everyone knows your name and they're glad you can't. So there's a lot we can do to make things more usable, but then also to make it more intelligent and more connected to, you know, to what people want. As these dynamic, multimodal products emerge, will natural language be enough to effectively interface with computers? So it could be the case that, like, over time, these things evolved into, like, you just speak natural languages, or do you think it will always be a component of a finite state
Starting point is 00:34:45 machine, a traditional computer, that's it? Yeah, I think this is right now an inflection point where we're sort of, you know, redefining how we interact with digital information. Yeah, yeah, yeah. And it's through, you know, the form of these AI systems that we collaborate with. And maybe we have several of them. and maybe they all have different competences. And maybe we have a general one that kind of follows us around everywhere,
Starting point is 00:35:13 knows everything about what my goals are, sort of in life, at work, and kind of guides me through and coaches me and so on. But, you know, there is also, we don't know exactly what the future looks like. And so we are trying to make these tools available and the technology available to a lot of other people. so they can experiment and we can see what happens. It is hard to imagine a world where AI doesn't continue to evolve and disrupt the world as we know it.
Starting point is 00:35:45 But as this happens, a common reaction is to wonder what happens to all the jobs. Here's Martine, the man who opened this episode, closing us out with an important reminder. There's always a question when you have market dislocations, like they're staring you in the face. You know what's coming, what happens to the jobs, what happens to people. there's something called Jevon's Paradox, and it's very simple. Jevon's Paradox says very simply, if the demand is elastic, it turns out like there's unlimited demand for compute, even if you drop the price, the demand will more than make up for it, normally far more than make up for it.
Starting point is 00:36:20 This is absolutely the case of the internet, right? So you get kind of more value, more productivity, et cetera. And I personally believe when it comes to creating any creative asset or any sort of kind of work automation, clearly the demand is elastic. I think the more that we make that, the more people consume. And so I think that we're very much looking forward to massive expanse of productivity, a lot of new jobs, a lot of new things. I think it's going to follow just like the microchip and just like the internet. Thank you so much for listening to part one of our coverage from AI Revolution. We really hope this gave you a glimpse into what may be to come, from scaling
Starting point is 00:36:57 laws to multi-modality, and we will be back in a few days with more key lessons from the event. including how AI is disrupting design, games, and entertainment, plus modern-day turn tests, AI alignment, and future opportunities. And as a reminder, if you would like to listen to all the talks in full today, you can head over to A66.com slash AI Revolution. We'll see you soon. If you liked this episode, if you made it this far, help us grow the show. Share with a friend, or if you're feeling really ambitious,
Starting point is 00:37:33 You can leave us a review at rate thispodcast.com slash A16c. You know, candidly, producing a podcast can sometimes feel like you're just talking into a void. And so if you did like this episode, if you liked any of our episodes, please let us know. We'll see you next time.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.