Behind The Tech with Kevin Scott - Bill Gates, Co-chair, Bill & Melinda Gates Foundation

Episode Date: March 21, 2023

Today, we have a special guest joining us on the podcast—Bill Gates. With the rapidly evolving AI landscape, including the release of products like OpenAI’s ChatGPT and the new Bing, it was the pe...rfect time to have Bill join to talk about this unique moment in the history of computing. In this episode, Kevin talks with Bill about the latest in AI research, including the release of GPT-4, how past technology revolutions have led us to where we are today, how AI is evolving his philanthropic work, his love of reading, and so much more!   Bill Gates | GatesNotes  Kevin Scott   Behind the Tech with Kevin Scott   Discover and follow other Microsoft podcasts at aka.ms/microsoft/podcasts.

Transcript
Discussion (0)
Starting point is 00:00:00 It was stunning. It was mind-blowing. After the biology questions, I had them type in, what do you say to a father with a sick child? It gave this very careful, excellent answer that was perhaps better than any of us in the room might have given. And so it's like, wow, what is the scope of this thing?
Starting point is 00:00:31 Because this is way better. Hi, everyone. Welcome to Behind the Tech. I'm your host, Kevin Scott, Chief Technology Officer for Microsoft. In this podcast, we're going to get behind the tech. We'll talk with some of the people who have made our modern tech world possible and understand what motivated them to create what they did. So join me to maybe learn a little bit about the history of computing and get a few behind-the-scenes insights into what's happening today. Stick around.
Starting point is 00:01:03 Hi, welcome to Behind the Tech. We have a great episode for you today with a really special guest. Bill Gates, who needs no introduction given the unbelievable impact that he's had on the world of technology and the world at large over the past several decades, has been working very closely with the teams at Microsoft and OpenAI over the past handful of months, helping us think through what the amazing revolution that we're experiencing right now
Starting point is 00:01:30 in AI means for OpenAI, Microsoft, all of our stakeholders, and for the world at large. I've learned so much from my conversations that I've had with Bill over these past months that I thought it might be a great thing to share just a tiny little glimpse of those conversations with all of you listeners today. So with that, let's introduce Bill and get a great conversation started. Thank you so much for doing the show today. And I just wanted to jump right in with maybe one of the more interesting things that's happened in the past few years in technology, which is GPT-4 and chat GPT and the work
Starting point is 00:02:14 that we've been doing together at Microsoft with OpenAI. By the time this podcast airs, OpenAI will have made their announcement to the world about GPT-4. But I want to sort of set the stage. The unveiling of the first instance of GPT-4 outside of OpenAI was actually to you last August at a dinner that you hosted with Reid and Sam Altman and Greg Brockman and Satya and a whole bunch of other folks. And the OpenAI folks had been very anxious about showing you this because your bar for AI had been really high.
Starting point is 00:03:06 I think it had been really helpful actually, the push that you had made on all of us for what acceptable high-ambition AI would look like. I wanted to ask you, what was that dinner like for you? What were your impressions? What had you been thinking before? What, if anything, changed in your mind after you had seen GPT-4?
Starting point is 00:03:34 AI has always been the holy grail of computer science. When I was young, Stanford Research had shaky the robot that was trying to pick things up and there were various logic systems that people were working on. The dream was always some reasoning capability. Overall progress in AI until machine learning came along was pretty modest. Even speech recognition was just barely until machine learning came along was pretty modest.
Starting point is 00:04:10 Even speech recognition was just barely reasonable. So we had that gigantic acceleration with machine learning, particularly in sort of sensory things, recognizing speech, recognizing pictures, and it was phenomenal. And it just kept getting better. And scale was part of that. But we were still missing anything that had to do with complex logic, with being able to, say, read a text and do what a human does, which is, quote, understand what's in that text.
Starting point is 00:04:47 So as Microsoft was doing more with OpenAI, I had a chance to go see them myself independently a number of times. They were doing a lot of text generation, they had a little robot arm. The early text generation still didn't seem to have a broad understanding. You know, like it could generate a sentence saying Joe's in Chicago and then two sentences later say Joe's in Seattle, which in its local probabilistic sense, you know, a human has a broad understanding of the world from both experience and reading that you understand that can't be. versions of GPT-4, I said to them, hey, if you can pass an advanced placement biology exam where
Starting point is 00:05:48 you take a question that's not part of the training set or a bunch of them and give fully reasoned answers, knowing that biology textbook is one of many things that's in that training corpus, then you will really get my attention because that would be a heck of a milestone. And so, you know, please work on that. I thought, you know, that would, they'd go away for two or three years because my intuition has always been that we needed to understand knowledge representation and symbolic reasoning in a more explicit way so that we were one or two inventions short of something where it was very good
Starting point is 00:06:31 at reading and writing and therefore being an assistant. And so it was amazing that, you know, you and Greg and Sam over the summer were saying, yeah, it might not be that long before we want to come demo this thing to you because it's actually doing pretty well on scientific learning. And in August, they said, yeah, let's schedule this thing. And so in early September, we had a pretty large group over to my house for a dinner, I think maybe 30 people in total, including a lot of the amazing opening eye people, but a good-sized group from Microsoft. Satya was there, and they gave it AP biology questions and let me give it AP biology questions. And with one exception, it did a super good job.
Starting point is 00:07:31 And the exception had to do with math, which we can get to that later. But it was stunning. It was mind-blowing. After the biology questions, I had them type in, what do you say to a father with a sick child? And it gave this very careful, excellent answer that, you know, was perhaps better than any of us in the room might have given. And so it's like, wow, what is the scope of this thing?
Starting point is 00:08:07 Because this is way better. And then the rest of the night, we'd ask historical questions about, are there criticisms of Churchill or different things? And it was just fascinating. And then over the next few months, as I was given an account and Sal Khan got one of those early accounts, the idea that you could have it write college applications or, you know, rewrite, say, the Declaration of Independence the way a famous person like Donald Trump might have written it. And it was so capable
Starting point is 00:08:46 of writing, you know, writing poems, you know, give it a tune like Hey Jude and tell it, you know, to write about that, tell it to take've said, wow, this is a fundamental change. Not without some things that still need to be worked out, but it is a fundamental advance. And it's confusing people in terms of, well, it can't yet do this. It can't do that. It's not perfect at this or that. But, hey, natural language is now the primary interface that we're going to use to describe things even to computers. And so it's a huge, huge advance. Yeah, so I want to, like, there's so many different things to talk about here. But, like, maybe the first one is to talk a little bit about what it's not good at.
Starting point is 00:09:53 Because the last thing that I think we want to do is give people the impression that it is an AGI, that it is perfect, that there isn't a lot of additional work that has to happen to improve it and make it better. You mentioned math as one of the things. And so I thought maybe let's talk a little bit about what you think needs to be better about these systems over time and where we need to focus our energy. Yeah. So there appears to be a general issue that knowledge of context when it's being asked, okay, I tell you something and I generate something. Humans understand, oh, I'm making up fantasy stuff here or I'm giving you advice that if it's wrong, you're going to buy the wrong stock or, you know, take the wrong drug. And so humans have a very deep context of what's going on. Even the AI's ability to know that you've switched context, like if you're asking it to tell jokes and then you ask it a serious question where a human would sort of see from your face or the nature of that change that, okay, we're not in that joking thing. It wants to keep telling jokes.
Starting point is 00:11:06 You almost have to do the reset sometimes to get it out of the, hey, whatever I bring up, just make jokes about it. And so I do think that sense of context, there's work. It also, in terms of how hard it should work on a problem, you know, when you and I see a math problem, we know, wow, I may have to apply simplification five or six times to get this into the right form. And so, you know, we're kind of looping through how we do these reductions, whereas today the reasoning is a sort of linear chain of descent through the levels. And if simplification needs run 10 times, it probably won't. So, you know, math is a very abstract type of reasoning.
Starting point is 00:11:57 And right now, I'd say that's the greatest weakness. You know, weirdly, it can solve lots of math problems. And there are some math problems where if you ask it to explain it in an abstract form, make essentially an equation or a program that matches the math problem, it does that perfectly. And you could pass that off to a normal solver. Whereas if you tell it to do the numeric work itself, it often makes mistakes. And it's very funny because sometimes it's very confident that, hey, or it'll say, oh, I mistyped. Well, in fact, there's not a typewriter anywhere in this scene. So the notion of mistyping is really very weird. So whether these current areas of weakness,
Starting point is 00:12:51 it's six months, a year, or two years before those largely get fixed. So we have a serious mode where it's not just making up URLs, and then we have a more fanciful mode. You know, there's some of that already being done largely through prompts and eventually through training. And, you know, training it for math, there may be some special training that needs to be done. But these problems I don't think are fundamental.
Starting point is 00:13:24 And, you know, there are people who think, oh, it's statistical, therefore it can never do X. That is nonsense. Every example they give of a specific thing it doesn't do, wait a few months, and it's very good. So characterizing how good it is, the people who say it's crummy are really wrong. And the people who think this is AGI, they're wrong. And those of us in the middle are just trying to make sure it gets applied in the right way. There's a lot of activities like helping somebody with their college application. What's my next step? What haven't I done? You know, I have the following symptoms that are, in fact, you know, far within the boundary of things that it can do quite effectively.
Starting point is 00:14:21 Yeah. Well, I want to talk a little bit about this notion of it being able to use tools to assist it in reasoning. And I'll give you an example from this weekend with my daughter. So my daughter had this assignment where she had this list of 115 vocabulary words, and she had written a 1,000-word essay, and her objective was to use as many of these vocabulary words as she reasonably could in this 1,000-word essay, which is sort of a ridiculous assignment on the surface, right? But she had written this essay, and she was going through this list trying to manually figure out what her tally was on this vocabulary list,
Starting point is 00:15:04 and it was boring. And she was like, all right, I want the shortcut here. Like, Dad, can you get me a chat GPT account? And can I just put this in there and it will do it for me? And we did it. And chat GPT, which is not running the GBT-4 model, but I don't think GBT4 would have gotten this right either. I didn't quite get it right, like it was not precise.
Starting point is 00:15:30 But the thing then that I got her to do with me is I was like, well, let's chat GBT, write a little Python program that can very precisely, I mean, this is a very simple intro CS problem here. the fact that the Python code for solving that problem was perfect and I got my solution immediately. Like, it's just amazing. And like my 14-year-old daughter who doesn't program understood everything that was going on. I don't know, like, if you reflected much over these past months about, because essentially when we are trying to solve a complicated math problem,
Starting point is 00:16:08 we've got a head full of cognitive tools that we pick up, these abstractions that you're talking about to help us break down very complicated problems into smaller less complicated problems that we can solve. I think it's a very interesting idea to think about how these systems will be able to do that with code. Yeah, it's so good at writing. That's just a mind-blowing thing.
Starting point is 00:16:39 But when you can use natural language, like say for a drawing program, that you want various objects and you want to change them in certain ways, sure, you still want the menus there to touch up the colors, but the primary interface for creating a from-scratch drawing will be language. And if you want a document summarized, that's something that can do extremely well. And so when you have large bodies of texts, when you have text creation problems, there was a chat GBT where a doctor who has to write to insurance companies
Starting point is 00:17:22 to explain why he thinks something should be covered, that's very complicated. And it was super helpful. Now, he was reading that letter over to make sure it was right. In CHAT GPT-4, the version 4 stuff, we took complex medical bills and we said, please explain this bill to me. What is this? And how does it relate to my insurance policy? And it was so incredibly helpful at being able to do that. Explaining concepts in a more simpler form, it's very, very helpful at that. And so there's going to be a lot of tasks where there's just huge increased productivity, including a lot of documents, payables, accounts receivables. Just take the health system alone. There's a lot of documents that now software will be able to characterize them very effectively. Yeah. So one of the other things that I wanted to chat with you about, like you have this really
Starting point is 00:18:33 unique perspective in your involvement in several of the big inflection points in technology. So for two of these inflections, you were either one of the primary architects of the inflection itself or one of the big leaders. So we wouldn't have the PC personal computing ecosystem without you, and you played a really substantial role in
Starting point is 00:19:01 getting the Internet available to everybody and making it a ubiquitous technology that everyone can benefit from. To me, this feels like another one of those moments where a lot of things are going to change. And I wonder what your advice might be to people who are thinking about like, oh, I have this new technology that's amazing that I can now use. How should they be thinking about how to use it? How should they be thinking about the urgency with which
Starting point is 00:19:34 they are pursuing these new ideas? How does that relate to how you thought about things in the PC era and the Internet era? Yeah, so the industry starts really small, where computers aren't personal. And then through the microprocessor and a bunch of companies, we get the personal computer, IBM, Apple. And Microsoft got to be very involved in the software. Even the basic interpreter on the Apple II, very obscure fact, was something that I did for Apple. And that idea that, wow, this is a tool that,
Starting point is 00:20:24 at least for editing documents, that you have to do all the writing, that was pretty amazing. And then connecting those up over the internet was amazing and then moving the computation into the mobile phone was absolutely amazing. So once you get the PC, the internet, the software industry, and the mobile phone, the digital world is changing huge, huge parts of our activities. I was just in India, you know, seeing this,
Starting point is 00:20:59 how they do payments digitally, even, you know, for government programs. It's an amazing application of that world to help people who never would have bank accounts because the fees are just too high. It's too complicated. And so we continue to benefit from that foundation. I do view this, the beginning of computers that read and write, as every bit as profound as any one of those steps. And a little bit surprising because robotics has gone a little slower than I would have expected. And I don't mean autonomous driving. that's a special case that's particularly hard because of the open-ended environment and the difficulty of safety and what safety bar people will bring to that. But even factories where you actually have a huge control over the environment of what's going on, and you can make sure that
Starting point is 00:21:59 no kids are running around anywhere near that factory. So, you know, a little bit people are saying, okay, you know, these guys can overpredict, which that's certainly correct. But here's a case where, you know, we underpredicted that natural language and it's the computer's ability to deal with that and how that affects white-collar jobs, including sales, service, helping a doctor think through what to put in your health record. That, I thought, was many years off. And so all the AI books, even when they talk about things that might get a lot more productive will turn out to be wrong. Because we're just at the start of this, you could almost call it a mania like the Internet mania.
Starting point is 00:22:52 Yeah. But the Internet mania all of its insanities and things that, I don't know, sock puppets or things where you look back and say, what were we thinking? It was a very profound tool that now we take for granted. And even just for scientific discovery during the pandemic, the utility of the immediate sharing that took place there was just phenomenal. And so this is as big a breakthrough,
Starting point is 00:23:25 a milestone as I've seen in the whole digital computer realm, which really starts when I'm quite young. Yeah. I'm going to say this to you, and I'm just interested in your reaction because you will always tell me when an idea is dumb. But one of the things that I've been thinking for the last handful of years is that one of the big changes that's happening because of
Starting point is 00:24:00 this technology is that for 180 years from the point that Ada Lovelace wrote the first program to harness the power of a digital machine up until today, the way that you get a digital machine to do work for you is you either have to be a skilled programmer, which is like a barrier to entry that's not easy, or you have to have a skilled programmer anticipate the needs of the user and to build a piece of software that you can then
Starting point is 00:24:33 use to get the machine to do something for you. This may be the point where we get that paradigm to change a little bit, where because you have this natural language interface and these AIs can write code and like they will be able to actuate a whole bunch of services and systems that, you know, we sort of give ordinary people the ability to get very complicated things done with machines without having to have like all of this expertise that you and I spent many, many years building? No, absolutely.
Starting point is 00:25:09 Every advance hopefully lowers the bar in terms of who can easily take advantage of it. The spreadsheet was an example of that because even though you still have to understand these formulas, you really didn't have to understand logic or symbols much at all. And it had the input and the output so closely connected in this grid structure that you didn't think about the separation of those two. And that's kind of limiting in a way to a super abstract thinker, but it was so powerful in terms of the directness. Oh, that didn't come out right.
Starting point is 00:25:47 Let me change it. Here, there's a whole class of programs of taking like corporate data and presenting it or doing complex queries against, okay, have there been any sales offices where we've had 20% of the headcount missing? And are sales results affected by that? You could just, now you don't have to go to the IT department and wait in line and have them tell you, oh, that's too hard or something. Most of these corporate learning things, whether it's a query or a report or even a simple
Starting point is 00:26:27 workflow where if something happens, you want to trigger an activity, the description in English will be the program. And when you want it to do something extra, you'll just pull up that English or whatever your language is in and type that in. There's a whole layer of query assistance and programming that will be accessible to any employee. The same thing is true of, okay, I'm somewhere in the college application process and I want to know, okay, what's my next step and what's the threshold for these things?
Starting point is 00:27:10 It's so opaque today. So empowering people to go directly and interact, that is the theme that this is trying to enable. I wonder what some of the things are that you are most excited about, just in terms of application of the technology to the things that you care about deeply from the foundation or your personal perspective. You care a lot about education, public health, climate,
Starting point is 00:27:48 and sustainable energy. You have all of these things that you're working on. Have you been thinking about how this technology impacts any of those things? Yeah, it's been fantastic that even going back to the fall, OpenA and Microsoft have engaged with people at the Gates Foundation thinking about the health stuff and the education stuff. In fact, Peter Lee is going to be publishing some of his thinking, which is somewhat focused on rich world health, but it's pretty obvious
Starting point is 00:28:27 how that work, in a sense, is even more amazing in health systems where you have so few doctors and getting advice of any kind is so incredibly difficult. And so it is incredible to look at saying, okay, can we have a personal tutor that helps you out? Can you, when you write something, if you're going to some amazing school, yes, the teacher may go line by line and give you feedback. But a lot of kids just don't get that much feedback on the writing. It looks like configured properly, this is a great tool to give you feedback on writing.
Starting point is 00:29:17 It's also ironic in a way that people are saying, what does it mean that can people cheat and turn in computer writing? Kind of like when calculators came along. Oh, my goodness, what are we going to do about adding and subtracting? And, of course, they did create contexts where you couldn't use the calculator. And we got through that without it being a huge problem. So I think education is the most interesting application. I think health is the second most interesting.
Starting point is 00:29:54 Obviously, there's commercial opportunity in sales and service-type things, and that'll happen. You don't need any uh foundation uh type engagement on that you know we've been we brainstormed a lot with sol con and uh it's you know looks very promising because a class size of 30 or 20 you can't give a student individual attention. You can't understand their motivation or, you know, what keeping them engaged. They might be ahead of the class. They might be behind. And it looks like in many subject areas, by having this and having dialogues and giving feedback, for the first time first time will succeed in helping education. Now, we have to admit, except for this sort of prosaic thing of looking up Wikipedia articles
Starting point is 00:30:52 or helping you type things and print them out nice, you know, the notion that computers were going to revolutionize education largely are still more in front of us than behind us. I mean, yes, some games draw kids in, but, you know, the average math score in the U.S. hasn't gone up much over the last 30 years. And so, you know, the people who do computers can't say, hey, we want credit for that. Credit for what? It's not a lot better than it was. Obviously, the computers didn't perform some miracle there. five to 10 years, we will think of learning and how you can be helped in your learning
Starting point is 00:31:47 in a very different way than just looking at material. Yeah. And I know you think about this as a global problem. My wife and I with our family foundation think about it as a local problem for disadvantaged kids. You know, like one of the common things that we see is that parent engagement makes a big difference in the educational outcomes for kids. And if you look at the children of immigrants, you know, in East San Jose or East Palo Alto here in the Silicon Valley,
Starting point is 00:32:20 like, you know, often the parents are working two, three jobs. Like they're so busy that like they have a hard time being engaged with their kids. And sometimes they don't speak English. And so, like, they don't even have the linguistic ability. And you can just imagine what a technology like this could do where it really doesn't care what language you speak. It can bridge that gap between the parent and the teacher, and it can be there to help the parent understand where the roadblocks are for the child and to even potentially get very personalized to the child's needs and sort of help them on the things that they're struggling with. I think it's really, really very exciting. Yeah, just the language barriers, we often forget about that. And that comes up in developing world. India has a lot of languages.
Starting point is 00:33:16 And I was at the Bangalore Research Lab as part of that trip. And they're taking these advanced technologies and trying to deal with the tail of languages. So that's not a huge barrier. One of the things that you said at the GPT-4 dinner at your house is that you had this experience early in Microsoft's history where you saw a demo that changed your way of thinking about how the personal computing industry was going to unfold and that caused you to pivot the direction of the company. I wonder if you might be willing to share that with everyone. Yeah. So Xerox had made lots of money on copying machines. They got out ahead.
Starting point is 00:34:07 Their patents were there. The Japanese competition hadn't come in. And so they created a research center out in Palo Alto, which was forever known by its acronym, Palo Alto Research Center Park. And at Park, they assembled an incredible set of talent. Bob Taylor and others were very good judges of talent. So you end up with Alan Kay, Charles Simone Lee, Butler Lampson. And I don't want to leave anybody out, but there's like a bunch of other people. And they create a graphics user interface machine.
Starting point is 00:34:49 They weren't the only ones. There were people over in Europe doing some of these things, but they combined it with a lot of things. They put it on a network. They got a laser printer. And Charles Simone was there programming this and did a word processor that used that very graphical bitmap screen and let you do things like fonts, you know, stuff we take for granted now. at night, and he demoed what he had done with this BravoWord processor, and then he printed on the laser printer a document of all the things that should be done if there were cheap, pervasive computers, and he and I brainstormed that, and he updated the document, and printed it again. And it just blew my mind. And the agenda for Microsoft came out of, you know, that's like in 1979 that I'm with him. And computers are still, you know, completely character mode. And, you know, so that's when the commitment to do software for the Mac emerges from Steve Jobs having a similar experience with Bob Belleville at Xerox built a very expensive machine called the Star that they only sold a few thousand of because people didn't think of word processing as something you would pay.
Starting point is 00:36:32 You had to come in really at the low end. So PCs with first character mode, but later graphics, word processing. So I hired Charles. Charles helped do Word and Excel and many of our great things. And eventually, you know, 15 years after Charles had shown me his thinking and we'd brainstormed, we largely achieved through Windows and Office on both Windows and Mac, we'd largely achieved that piece of paper. So I told the group that that was the other demo that had kind of blown my mind and made me think, okay, what can be achieved in the next five to ten years? We should be more ambitious taking advantage of this breakthrough,
Starting point is 00:37:30 even with the imperfections that, you know, we're going to reduce over time. Yeah, it was a really powerful and motivating anecdote that you shared. So, you know, maybe one last thing here, you know, before we go, or maybe two more things. So what do you think are the, like, the big grand challenges that we ought to be thinking about, the next five to 10-year period. So in a sense, I actually have this piece of paper that Charles wrote. It's here by my desk frame because I think it was one of the more unbelievable predictions of a technology cycle that anybody's ever written. And I don't know why everybody doesn't know about the existence of this thing.
Starting point is 00:38:24 It's just unbelievable. But as you think about what lies ahead of us over the next five to 10 years, what's your challenge, not just to Microsoft, but to everybody in the world who's going to be thinking about this? What do you think we ought to go push on really, really hard? Well, there'll be a set of innovations on how you execute these algorithms. Lots of chips, some movement from silicon to
Starting point is 00:38:57 optics to reduce the energy and the cost. Immense innovation where NVIDIA is the leader today, but others will try and challenge them as they keep getting better and, you know, using even some radical approaches because we want to get the cost of execution on these things and even the training dramatically less than it is today. Ideally, we'd like to move them so that often you can do them on a self-contained client device, not have to go up to the cloud to get these things. So lots to be done on the platform that this uses.
Starting point is 00:39:41 Then we have an immense challenge in the software side of figuring out, okay, do you just have many specialized versions of this thing, or do you have one that just keeps getting better? And, you know, there'll be immense competition. Those two approaches, you know, even at Microsoft, we'll pursue both in parallel with each other. Ideally, within a contained domain, we'll get something that the accuracy is provably extremely high by limiting the areas that it works in and by having the training data
Starting point is 00:40:25 and even perhaps some pre-checking, post-checking type logic that applies to that. I definitely think areas like sales and service, that there is a lot that can be done there and that that's super valuable. The notion that there is this emergent capability means that the push to try and scale up even higher, that'll be there. Now, what corpuses exist? Once you get past every piece of text and video, are you synthetically generating things? And do you still see that, you know,
Starting point is 00:41:07 improvement as you scale up? Obviously, that'll get pursued. And, you know, the fact that it costs billions of dollars to do that won't stand in the way of that going ahead in a very high-speed way. And then, you know, there's a lot of societal things of, okay, where can we push it in education? It's not that it'll just immediately understand student motivation or student cognition, there'll have to be a lot of training and embedding it in an environment where the adults are seeing the engagement of the student and seeing the motivation. And so even though you free up the teacher from a lot of things, that personal relationship piece, you're still going to want all the context coming out from that those tutoring sessions to be visible and help
Starting point is 00:42:15 the dialogue that's there with the teacher or with the patient. Microsoft talks about making humans more productive. So some things will be automated, but many things will just be facilitated where the final engagement is very much a human, but a human who's able to get a lot more done than ever before. So the number of challenges and opportunities created by this is pretty incredible. And I can see how engaged the OpenAI team is by this.
Starting point is 00:43:01 I'm sure there's lots of teams I don't get to see that are pushing on this. And the size of the industry, I mean, when the microprocessors invented, the software industry was a tiny industry. I mean, we could put most of us on a panel, and they could complain that I work too hard and it shouldn't be allowed. We could all laugh about that. You know, now this is a global industry with, and so it's a little harder to get your hands around. I get a weekly digest of all the different articles
Starting point is 00:43:37 about AI that are being written. You know, oh, can we use it for moral questions, which seems silly to even ask me, but fine, people can ask whatever they want to ask. And this thing has the ability to move faster because the amount of people and resources and companies is way beyond those other breakthroughs that I brought up and was privileged to live through. Yeah.
Starting point is 00:44:13 I mean, one of the things for me that has been really fascinating, and I think I'm going to say this just as a reminder to folks who are thinking about pursuing careers in computer science and becoming programmers. So I spent most of my training as a computer scientist in my early part of my career as a systems person. So I wrote compilers and tons of assembly language and design programming languages, which I know you did as well. I feel like a lot of the things that I studied just in terms of parallel optimization and high-performance computer architectures in grad school,
Starting point is 00:44:58 I left grad school and went to Google and thought I would never use any of this stuff ever again. Then all of a sudden now, like we're building supercomputers to train models and these things are relevant again. And I think, you know, it's interesting. Like, I wonder what Bill Gates, the young programmer, would be working on, you know, if you were in the mix right now, like writing code for these things, because there's so many interesting
Starting point is 00:45:27 things to go work on. What do you think you as a 20-something-year-old young programmer would get really excited about in this stack of technologies? Well, there is an element of this that's fairly mathematical. So I feel lucky that I did a lot of math, and that was a gateway to programming for me, including all sorts of crazy stuff with numerical matrices and their properties. And so there are people who came to programming without that math background who do need to go and get a little bit of the math. I'm not saying it's super hard, but they should go and do it so that when you see those funny equations, you're like, oh, okay, I'm comfortable with that
Starting point is 00:46:16 because a lot of the computation will be that kind of thing instead of classic programming. It is the paradox. When I started out, writing tiny programs was super necessary. Like, you know, the original Macintosh is a 128K machine, 128K bytes, 22K of which is the bitmap screen. And so almost nobody could write programs that fit in there. And so Microsoft, our approach, our tools, let us write code for that machine. And really only we and Apple succeeded.
Starting point is 00:46:58 Then when it became 512K, a few people succeeded. But even that, people found very difficult. And I remember thinking, you know, as memory got to be, you know, for gigabytes, all these programmers, they don't understand discipline and optimization. And, you know, they're just allowed to waste resources. But now that, you know, these things that you're operating with billions of parameters, the idea of, okay, do I, can I skip some of those parameters? Can I simplify some of those parameters?
Starting point is 00:47:31 Can I pre-compute various things? You know, if I have many, many models, can I keep deltas between models instead of having them? All the kind of optimizations that made sense on these very resource-limited machines, well, some of them come back in this world where when you're going to do billions of operations or literally hundreds of billions of operations, you know, we are pushing the absolute limit of the cost and performance of these computers. And that's one thing that is very impressive is the speedups, even in the last six months on some of these things, has been better than I expected. And that's fantastic because you get the hardware speed up, the software speed up, kind of multiplied together.
Starting point is 00:48:28 That means how resource bottlenecked will we be over the next couple of years? That's less clear to me now that these improvements are taking place, although I still worry about that and how we make sure that companies broadly, and Microsoft in particular, allocates that in a smart way. Understanding algorithms, understanding why certain things are fast and slow, that is fun. The systems work that in my early career is just one computer and later kind of a network of computers.
Starting point is 00:49:08 Now that systems where you have data centers with millions of CPUs, it's incredible the optimization that can be done there. Just how the power supplies work or how the network connections work. Anyway, in almost every area of computer science, including database-type techniques, programming techniques, this kind of forces us to think about in a really new way. Yeah, I could not agree more. So last, last question. I know that you are incredibly busy and you have the ability to choose to work on whatever it is that you want to work on. But I want to ask you anyway, like, what do you do outside of work for fun? I ask everybody who comes on the show that. Well, that's great. I get to read a lot.
Starting point is 00:50:06 I get to play tennis a lot. During the pandemic, I was down in California in the fall and winter, and I'm still enjoying that, although the foundations meeting in person and some of these Microsoft OpenAI meetings, it's been great that we've been able to do those in person. But some we can just do virtually. So anyway, I play pickleball because I've been playing for over 50 years. Tennis. And I like to read a lot. I goofed off and went to the Australian Open for the first time because it's nice warm weather down there. So I actually want to push on this idea of read a lot. and open for the first time because it's nice warm weather down there and not.
Starting point is 00:50:49 So I actually want to push on this idea of read a lot. So you say you read a lot, which is not the same as what most people say when they say they read a lot. You're famous for carrying around a giant tote bag of books with you everywhere you go and you read an insane amount of stuff. You have everything from really difficult science books, you know, all the way to fiction. So like, how much do you actually read? Like what's a typical pace of reading for Bill Gates? You know, if I don't read a book in a week, then I'm really, I can re-look at what I was doing that week. If I'm on vacation, then I'll, you know, hope to read more like five, six, or seven. Of course, books are quite variable in size.
Starting point is 00:51:34 You know, over the course of the year, you know, I should be younger children who read even more than I do. So it's kind of like, oh, geez, you know, I have to be, you know, which solo books am I going to read? I still, you know, read all the Schmiel, Pinker, youer, some authors that are just so profound and have shaped my thinking. But reading is kind of relaxing. I should read more fiction. When I fall behind, my nonfiction list tends to dominate. And yet, people have suggested such good fiction stuff to me. And that's why I kind of share my reading list on Gates Notes.
Starting point is 00:52:28 So what's the over-under? You're famous for saying that you want to read David Foster Wallace's Infinite Jest, like over-under that happening in 23. Well, if there hadn't been this darn AI advance that's distracting me, I'm kidding you. But it's really true. I have allocated, and with super excitement, a lot more time to sitting with Microsoft product groups and saying, okay, what does this mean for security? What does it mean for Office? What does it mean for security? What does it mean for office? What does it mean for our database
Starting point is 00:53:05 typing? Because I love that type of brainstorming because new vistas are opened up. So no, it's all your fault. No infinite jest this year. Excellent. Well, I appreciate you making that trade because it's been really fantastic over the past six months having you help us think through all of this stuff. Thank you for that and thank you for doing the podcast today. Really, really appreciate it. No, it's fun, Kevin. Thanks so much. Thank you so much to Bill for being with us here today.
Starting point is 00:53:40 I hope you all enjoyed the conversation as much as I did. There's so many great things about this conversation, I hope you all enjoyed the conversation as much as I did. There's so many great things about this conversation, which were reflections of the conversations that we've been having with Bill over the past handful of months as we think about what AI means for the future, is how he thought about what personal computing and the microprocessor and PC operating system revolution meant for the world when he was building Microsoft from the ground up. Even what it felt like for him as one of the leaders helping bring the internet revolution to the world.
Starting point is 00:54:28 So like those parts of the conversation today that we had where he was recounting some of his experiences, like the first meeting that he had with Charles Simoni at Xerox PARC, where he saw one of the world's first graphical word processors and how seeing that and talking with Charles influenced an enormous amount of the history of not just Microsoft, but the world in the subsequent years. Just hearing Bill talk about his passion for the things that the Gates Foundation is doing and what these AI technologies mean for those things, like how maybe we can use
Starting point is 00:55:08 these technologies to accelerate some of the benefits to the people in the world who are most in need of technologies like this to help them live better and more successful lives. Again, this is a tremendous treat for being able to talk with Bill on the podcast today. If you have anything you'd like to share with us, please e-mail anytime at BehindTheTechatMicrosoft.com.
Starting point is 00:55:35 You can follow Behind the Tech on your favorite podcast platform or check out full video episodes on YouTube. Thanks for tuning in and see you next time.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.