ACM ByteCast - Peter Lee - Episode 68

Episode Date: April 17, 2025

In this episode of ACM ByteCast, our special guest host Scott Hanselman (of The Hanselminutes Podcast) welcomes ACM Fellow Peter Lee, President of Microsoft Research. As leader of Microsoft Research, ...Peter incubates new research-powered products and lines of business in areas such as AI, computing foundations, health, and life sciences. Before Microsoft, he established a new technology office that created operational capabilities in ML, data science, and computational social science at DARPA, and before that he was head of the CS department at CMU. Peter served on President Obama’s Commission on Enhancing National Cybersecurity and has testified before both the US House Science and Technology Committee and the US Senate Commerce Committee. He coauthored the bestselling book The AI Revolution in Medicine: GPT-4 and Beyond. In 2024, he was named by Time magazine as one of the 100 most influential people in health and life sciences. In the interview, Peter reflects on his 40+ years in computer science, from working on PDP-11s and Commodore Amigas to modern AI advancements. He highlights how modern technologies, built on decades of research, have become indispensable. He also talks about his healthcare journey, including work that earned him election to the National Academy of Medicine, and the potential (and limitations) of AI in medicine. Peter and Scott touch on the impact of LLMs, the lack of ethics education in traditional CS curricula, the challenges posed by growing AI complexity. Peter also highlights some important Microsoft Research work in AI for Science and Quantum Computing.

Transcript
Discussion (0)
Starting point is 00:00:00 This is ACM Bytecast, a podcast series from the Association for Computing Machinery, the world's largest education and scientific computing society. We talk to researchers, practitioners, and innovators who are at the intersection of computing research and practice. They share their experiences, the lessons they've learned, and their own visions for the future of computing. I'm your host today, Scott Hanselman. Hi, I'm Scott Hanselman. Hi, I'm Scott Hanselman.
Starting point is 00:00:26 This is another episode of Hansel Minutes, an association with the ACM Bytcast. Today, I'm chatting with Dr. Peter Lee, President of Microsoft Research. You've got a resume as long as my arm, and it is an absolute joy to chat with you today, sir. It's a joy for me to be here. I really appreciate it. I want to go in a little bit of a controversial direction because I want to start out with the fact that I had a birthday last week.
Starting point is 00:00:49 And in this year, if I understand correctly, you're going to turn 65. Is that correct, sir? Yes, which I can't believe. Yeah. Yeah. I turned 51 last week and I have been reflecting on the space that I've been in, in the last 32 years of software experience and you have more. What does that feel like? Because I've got, if you can see behind me, I've got a PDP-11 that I built from a Raspberry Pi, and I've been learning about that stuff. I've got my Commodore 64 that I had as a child.
Starting point is 00:01:22 I'm really reflecting on how far this has come. With the experience and the space that you have occupied in computer science over the last 40 plus years, what does that feel like as we sit here on this AI moment? Yeah. First off, since you mentioned PDB-11s and Commodore Amigas, there's a lot of fondness in my heart for those. I actually, my first real paying job was assistant administrator for a PDP 11.
Starting point is 00:01:52 And so I remember learning how to wire wrap the core memories there. And then for Commodore, Amiga, Amiga NUT 64, took time away from my graduate studies at Michigan to be part of a startup. In that startup, one thing we were trying to do was put productivity software, like word processing and spreadsheets and paint programs, onto Apple IIe and Commodore Amiga. So if you ever used Amigarrite, it has MyCode. Really? And that company didn't really succeed, but we had to do things like make our own mouse,
Starting point is 00:02:33 because of course those computers didn't have mice at that time, but you needed a mouse to use those sorts of modern productivity programs. And so really a lot of fondness for those. I reflected about this and Scott, I assume you have a mobile phone in your pocket. Of course. The reason I think you have that is you look pretty comfortable and relaxed because I think nowadays, if you forgot it at home or left it in a hotel room, you feel bad, you feel naked and
Starting point is 00:03:04 vulnerable like you can't really function properly. And from a research perspective, I count six major triumphs of computer science research that are in that mobile phone. There's VLSI design that emerged in the 1970s. There's a Linux or Unix inspired operating system kernel. There is a software defined mobile wireless radio and so on. And these things were hardcore academic research in computer science departments, great ones around the world, in great laboratories,
Starting point is 00:03:36 you know, like Bell Labs and others. They got published and then they made it into a suite of technologies that you literally feel like you can't live without all day every day. And you know it's kind of amazing when you reflect on things like that. And so when I look at for example what's happening today with LLMs in generative AI, the question in my mind isn't one about AGI or not. It's about when and if say large language model technology will become yet another example of a technology that you can't bear to be without at any point of your waking hours.
Starting point is 00:04:18 And I think it might happen, but you know, just, you know, as I've get closer to retirement at some point and just reflect on everything that's happened, it's really incredible how much things have evolved to the point where we really literally depend on them all the time for our mental health and sanity. Yeah. That idea of not being able to live without it is such an interesting way of phrasing it. I have a very personal relationship with my phone
Starting point is 00:04:46 because it also runs my artificial pancreas. And I was in Johannesburg last week for two weeks and I bought a backup phone because even though I can drive stick shift and just take needles and take blood sugar, you know, with my finger, like I could do that. And I did that for 20 years. Why is it now that I feel like I cannot survive without it?
Starting point is 00:05:09 I have to have the phone within Bluetooth range, you know, near me 24 hours a day. And if I lost the phone or if it was dropped or stolen, I would immediately need to hook another phone up because I want my continuous glucose meter. And my wife, who's not diabetic but is a nurse, said, well, you could just take needles and stick your finger like you did back in the 90s. And I was like, that's unthinkable. How dare you? That's actually something called the flip.
Starting point is 00:05:34 And you see the flip most vividly in the technology adoption history in healthcare and medicine. And my favorite example of that is ultrasound. So, you know, ultrasound had been invented quite a long time ago, but it wasn't until the late 1950s that technologists proposed that ultrasound might be a good diagnostic tool in the practice of medicine. And so, you know, if you're a pregnant woman and you go to see your gynecologist or obstetrician, you will see ultrasound used to
Starting point is 00:06:05 examine your unborn fetus and the reproductive system of the mother. But in the late 1950s, when this was first proposed, it scared people. Is this going to damage my baby? Is it going to damage my ability to reproduce? Are doctors trained to read these grainy images and all sorts of problems.
Starting point is 00:06:26 And so it took about a decade before ultrasound became the standard of care. And so now, what do I mean by the flip? Well today, if you are a pregnant woman and you go to your obstetrician and your obstetrician says, ah, I don't believe in that fancy ultrasound technology. I'm going to put my hands on your belly and use manual palpation." Not only would you be horrified at that, but you might even report that as malpractice. And so expectations, not only in the standard of care, but in patient demands and expectations, consistently flip in the history of technology adoption medicine.
Starting point is 00:07:03 And your example is another example of how things flip. You flip from being able to drive manually and monitor your own blood sugar levels to a feeling that why would you do that? It's so primitive, it's so inexact, so dangerous, unconscionable, and in fact irresponsible to do it that way. I'm curious though that I find myself feeling like old man who shakes fist at cloud because I feel like the young people are saying it's going to change the world and if you're okay,
Starting point is 00:07:34 boomer, if you're against it, then if you're not with us, you're against us. But I remember when the TI-83 and the TI-81 calculators came out and the math teachers were like, no one's going to be able to do math in their head anymore because you won't always have a calculator young man, and here we are with a pocket supercomputer. I am finding myself questioning my own opinions about tech based on my age and my generation and I'm really trying to be introspective. I get that it's going to change the world,
Starting point is 00:08:02 but I also think that cognitively, we don't know how these's going to change the world, but I also think that cognitively, we don't know how these are going to change our brains. And I don't know if having a pocket supercomputer has changed our brains for the positive. Yeah. You know, I'm always a little bit more circumspect. It's easy as two people who are tech executives, you know, to look at all the historical examples of, you know, so the adoption of advanced technologies that are really enabling,
Starting point is 00:08:29 there is a consistent pattern. And so you could always argue, as we in tech do, that the right side of history is that these things advance and the world gets better and better. And I think that's generally true, but I think it's also true, and it's worth some reflection that we lose some things along the way. And that isn't to say that the world isn't getting better and life isn't getting better because I think it is by any measure. But I think it's wrong to deny
Starting point is 00:08:58 that we're losing some things and some skills and some abilities that maybe were important. One thing I would say about, especially in this new AI era is, I do see that the world's leading thinkers are actually earnestly trying to be thoughtful about this. That there is a big debate. In fact, it's a debate of leading thinkers that I haven't seen since the human genome was mapped. When we finished mapping human genome
Starting point is 00:09:25 as a scientific community, it sparked a huge debate about what would this mean? And would genetic engineering be a good or a bad thing? And the fact that there were so many leading thinkers and the whole academic and research area of bioethics really just mushroomed into hugely important kinds of thought leadership and research. I think that's helped us a lot to gain as much of the goodness out of our understanding
Starting point is 00:09:56 of genetics while mitigating the potential downstream harms and risks. And I see the same intensity of debate going on with AI as people are trying to grapple with what this means. I think that that's a good thing and it actually makes it easier for me to be optimistic about what's happening right now. Yeah. As someone who is and was a professor of computer science for many, many years, did you plug ethics in and
Starting point is 00:10:22 did you have people thinking about the tech? I mean, sometimes it's fun to just do techie stuff for techie reasons, but I've always come back to, are we making someone's life better by doing this or are we making someone's life demonstrably worse? But a lot of people, when I do these informal polls, I say, all right, who here has ever taken a computer science ethics class? The hands are not going up, a majority or plurality. It feels like we're not teaching computer science ethics and it feels like it's even more important in a post-GPT-4 world.
Starting point is 00:10:51 Yeah, I have to admit, earlier in my career, it really wasn't a serious thought. In fact, if anything, I tried hard to adopt the Silicon Valley stereotype ethos of tech is good and more tech is better. And just the belief that technology could be the answer to all of our problems and that we can solve all problems through technology. I do think we've gotten to such a more enlightened state about this. I think a lot of that has come through lessons learned
Starting point is 00:11:25 the hard way. We've seen so clearly that tech isn't different than any other technology in being dual edged. And in fact, it's more interesting and has greater positive and negative potential because information technology can be democratized. It can be given to literally everyone. You know, unlike genetic engineering where you have to have a multi-million dollar wet lab to do anything here, literally every person on the planet could gain access and harness this for both good and evil. And so I think ethics has become a much bigger deal. Whereas I think earlier in my career, personally, I think it was an age of innocence, where you only saw the goodness in things.
Starting point is 00:12:06 Yeah. I really appreciate the democratization of it. Open source itself, the pancreas that I use is a toolkit that one could build from scratch, and I did, and that anyone could go and do that. Now we're seeing open weights and open models. What's your thinking about frontier models that are closed versus frontier models that are more open
Starting point is 00:12:25 about where they came from, what they were trained on, and the corpus that built them up? Yeah, I think that this is also evolving really quickly. The thing that's so interesting to me is the cost of doing something first versus the cost of doing something versus the cost of doing something tenth. And so if you are OpenAI or one of OpenAI's frontier competitors, you're trying to get to certain levels of intelligent capability first. And there you can see a pretty methodical investment strategy. To get to GPT-2 maybe requires on the order of $5 to $10 million of compute cost to train a model at that level.
Starting point is 00:13:10 And GPT-2 allows you to see perfect loss curves. So that gives you confidence that further scaling is going to get you somewhere. And then to get to GPT-3 level and to do that first, pretty consistently in whatever domain you're trying to train your A models consistently requires about 10x the compute cost. So now you're talking 50 to 100 million dollars and then to get to GPT-4 level you know whether it's in language or molecular dynamics or you know weather modeling whatever seems to take yet another 10x leap if you want to get there first. So now, you know, you're starting to get close to a billion dollars and so on.
Starting point is 00:13:50 The thing that's so interesting is, like in all things that have to do with technological invention, once smart people see that something is possible, then it becomes so much easier for the second and third and fourth and tenth people to do it. And what we're seeing in the industry is that not being first means that you can also, of course, you don't benefit from being first, but you benefit from being able to be smarter and a hell of a lot cheaper to get to the same state. Not that getting to a GPT-4 class base model is cheap. It's still quite expensive, but it's a lot cheaper than it took for the very first innovators to get there. This is the pattern of tech that you and I have lived through
Starting point is 00:14:35 our entire careers. So we shouldn't be surprised by this, but just to see it playing out with just this stunning speed is just incredible. And business-wise, the question is, there's always been a big premium in our business in being first. Now the question is, given the rapid pace of change and evolution, is there still the same amount of value in being first as there has always been?
Starting point is 00:14:57 The idea that we shouldn't be surprised, but we consistently are surprised. I remember the first time that I saw, you know, a chat bot for lack of better words. And now like I'm looking at my phone, I've got an app called MLC chat and I've got five three, five mini it's a, a four bit quantized F 16, but I'm having a conversation in airplane mode on a, you
Starting point is 00:15:21 know, on a 15 tops iPhone. That seems insane to me and I get it and I understand it. But we're at a point now where the full stack, you know, you always hear about the full stack engineer is so deep that I find myself going back to building things from original parts, just so I can remind myself, like I've got a, an Apple one that I'm building from 7,400 series chips,
Starting point is 00:15:45 just to kind of get my hands back in the silicon because the stack is so high, I'm trying to get my hands dirty so that I can emotionally like accept that there's an SLM on my phone that works in airplane mode. You know, by the way, I'll take one of those Apple Ones
Starting point is 00:16:03 if you make two of them. Yeah, yeah, those are great. That's a company called Smarty Kit. I'll send that to you. You know, there's something that you're saying there though that is another reflection just looking back at my career. And I think you and I, when we started in all of this, we were able to wrap our heads around and know the intricate details of the entire stack end to end. From the silicon, or in my case, the wire wraps,
Starting point is 00:16:29 all the way to all the code, every line of code in the applications. And things are just so much more complicated now. And when we think about what AI is going to enable, I think AI is very quickly going to enable us to construct systems that are so complex that they will really completely defy any ability for humans to comprehend them fully. In research, where I think this might happen first, Right now, there's a vibrant set of researchers that are using generative AI to write mathematical proofs. And when you ask an AI system to write a mathematical proof,
Starting point is 00:17:13 you generally ask it to write it in a proof language, a popular one is called Lean. There are others as well. The interesting thing about a proof language is that they're set up so that you can use a simple type checker, just like you would have in any programming language. And these proof languages are set up so that if they type check, you know for certain that the proof is valid. And so I foresee in much less than five years that we'll have an AI system generate a proof of some mathematical theorem. We'll be able to type check that proof to know that that proof is absolutely valid and correct. But that proof itself might defy any ability for any human being, even the world's smartest human beings, to understand it. And that's sort of like a point example of AI getting us to a place where we're able to construct things, build things,
Starting point is 00:18:11 that work. And we know we can see them work, we can validate that they work, but we won't know how or why. And I think it's going to be, when people ask me about AI or superintelligence, that will be the first mark of it for me. And, you know, your roots and my roots, we still want to, you know, understand everything. And we do things to keep that fresh for us. But I think it's going to get hard. ACM Bytecast is available on Apple Podcasts, Google Podcasts, Podbean, Spotify, Stitcher, and TuneIn. If you're enjoying this episode,
Starting point is 00:18:50 please do subscribe and leave us a review on your favorite platform. That's a really interesting point. I think you're right that I and perhaps people of our generation plus or minus a few years are still unwilling to let go of the fact that we took a rock, we flattened it, we infused it with lightning, and now it talks to us and I want to understand the lightning and the squishy rock part.
Starting point is 00:19:14 But my children, my 19-year-old, is perfectly willing to just accept that the magic black box is doing a thing. When I have to have conversations with him about maybe, let's not anthropomorphize the AI, let's talk about what's really happening. He's like, it helped me with my homework and I'm cool with that. I was a little bit taken by your book,
Starting point is 00:19:34 The AI Revolution in Medicine that was pre-GPT-4 general availability. You were poking at the model in ways where you're like, you were anthropomorphizing it, but you were also trying to understand it. What would you change about how you interacted with the model now that you know more, since this very good book has been published? Yeah, you know, that book, you know, we wrote it over the Christmas holidays in 2022, you know, while GPT-4 was still a secret project. And we organized the ourselves so that the book would get published at the same month that GPT-4
Starting point is 00:20:12 was released to the whole world. And so that was a time when, you know, we were just so amazed and baffled. You know, I talk about this as the nine stages of grief. In fact, I think at some point in the book, I talk about the nine stages of AI grief. When GPT-4 was first exposed to me by the folks at OpenAI, I was intensely skeptical because they were just claiming that this thing could do things that I just thought were not possible. And then you pass from that stage of skepticism to the stage of frustration, because I felt like I was seeing my colleagues in Microsoft research falling, getting duped by this thing.
Starting point is 00:20:53 And then you start to feel worried because I detected that, wow, Microsoft might actually make a big bet on this thing. But then you get hands-on and you start to encounter things that are just amazing. And I remember feeling the joy that, wow, this thing is, I never thought I would live long enough to see such a technology little and have my hands on it. And then you get into a period of intensity. So there are these stages that you go through, but in those early stages of joy and euphoria and you lose sleep, it is those aspects that feel empathetic that draw you into anthropomorphism that end up being so interesting. And in the field of medicine, this has been observed over and over again.
Starting point is 00:21:39 In fact, one month after we published our book, UC San Diego and Stanford jointly published a paper in a medical journal where they used GPT-4 to respond to emails from patients. They compared those to the emails that doctors, human doctors, would write. Then they had a blind test and they had patients and doctors grade the quality, correctness of these responses to patient queries. Not only was the AI equally accurate, but by a factor of nine to one, the AI-generated notes were judged by patients to be more empathetic.
Starting point is 00:22:17 And of course, you know, it seems crazy to say a machine can be empathetic. What it really means, I think, is that a frazzled doctor can't take the time to write more than two or three sentences and just get to the point and then get onto the next email. Whereas the AI can write a couple of paragraphs and might remember that during the encounter they were talking about going to a Seahawks game and other stuff like that and put in those nice personal touches.
Starting point is 00:22:46 And so there's something there that is both interesting and disturbing, but also seems to really touch people in a very meaningful and very practical way. And I think we're still, just as a we're trying to come to grips with this. Yeah. I'm struck that one could theoretically say that it can be infinitely empathetic and infinitely patient given appropriate prompting and an appropriate good attitude on the part of the controlling human. I keep coming back to empathy. We need more empathy in the world right now,
Starting point is 00:23:22 we need more empathy in tech. When I use Copilot on GitHub Copilot, I don't ask it how to do my homework. I basically use it as an enthusiastic pair programming partner and I find it to be infinitely patient. It never judges me, it's never mean, it's never unkind. So then it has theoretically infinite empathy if I talk to it right. I'm struck by the epilogue in your book. You say, it just can't be that
Starting point is 00:23:49 next word prediction could be intelligence or can it? Am I just a statistical model of what's the most likely next word from Hansel and to say? Is that the animus of me? I think, yeah, at the time we wrote that book, you know, this was, I was really baffled by this, but I think my understanding and acceptance of what's going on has evolved a lot more. It's true that the fundamental pre-training of these large language
Starting point is 00:24:19 models is to predict the next word in the conversation. And so to that extent, you could say that these language models aren't trained to do anything useful except that. But here's an example that I like to give. Let's take the sentence, and the murderer is blank. Okay, so now if you want to, you know, pick with the highest quality, the word that would fill in the blank, well
Starting point is 00:24:48 that sentence, and the murderer is blank, is in the context of let's say a whole Agatha Christie murder mystery novel. And if you were to just approach this purely as a statistics question, well there are thousands of murder mystery novels and short stories, and you could try to make a statistical pattern on what are the most likely names of murderers, and you wouldn't get any good answer there. Instead, in order to really optimize the quality of the fill-in-the-blank capability, you somehow have to be able to do some deductive reasoning. You have to have an understanding of the psychology of humans in different situations,
Starting point is 00:25:29 what motivates them, how they react under certain kinds of questioning, and all of those things. And so the way to understand what's going on when we try to optimize, fill in the blank or next word prediction, is as a very, very large, astronomically large stochastic process that has a chance of accidentally discovering neural circuitry that implements some aspects of those reasoning functions. And the fact that that can happen at all, even by accident, is amazing. But we're operating a scale where, indeed, it is actually happening. And so it's not that it's next word prediction that is causing us to appear
Starting point is 00:26:15 to be thinking, but the process of highly optimizing an ability to do very, very good next word prediction is giving us a chance to really discover and solidify these bits of neural circuitry to do things. Yeah. One analogy that I've used to explain to some young people, and I don't know if it's a good one, is that we are, as humans, limited by the size of our stack for our recency. And then the older you get, you get this larger and larger and larger heap that you can pull from. And then you're constantly pulling things out of the heap and into the stack.
Starting point is 00:26:48 And then there's certain presenters, there's certain thinkers. Jamel Bouie is one I think of who's an opinion person for the New York Times, who seems to have this huge corpus of information that he is pulling upon all the books that he's ever read. He's had an amazing vocabulary. And I admire him and his ability to page in and out these pieces of wisdom while I struggle to root around in the totality of my existence. I think that AI will feel like AGI when its stack,
Starting point is 00:27:19 when its context window is beyond those of even the smartest person, and that's going to be the thing, that it has just such a large context window, it's bigger even the smartest person. That's going to be the thing, that it has just such a large context window, it's bigger than a human lifetime. Yeah. One thing that up till now, for the most part, large language models and transformers specifically were pretty imperfect
Starting point is 00:27:41 in memorization of things. There's a massive compression of the training corpus that goes on when you train it into a transformer. And so, you know, I was always trying to explain that to doctors as they're trying to come to grips with generative AI because, you know, unlike a normal computer, in fact, one popular application, maybe the most popular application that doctors use online is something called UpToDate. It's essentially a search engine for very highly curated medical knowledge. And so in UpToDate, you ask a question and you get a
Starting point is 00:28:20 medically precise answer. And we've trained ourselves to use that just like we use web search. You make a query and you expect to get a set of answers that are pretty precise. But the transformer doesn't have that capability. It has very, very imperfect memory. Of course, that's now evolving because there's no fundamental reason why computer-based system couldn't actually have perfect memory recall. And so I agree with you. I think we're going to get to a point where these AI systems really are going to be benefiting from the fundamental capabilities
Starting point is 00:28:58 of perfect memory recall that we've always assumed machines would have, while also making all of these associations and engaging in this reasoning. Why is it, and forgive my ignorance if this is a question I should know the answer to, but when did your deep interest in the medical aspect of technology, like in everything that you do, and certainly recently of late,
Starting point is 00:29:19 you always want to make technology to make people's lives better, and it seems to come back to healthcare. Do you have a background in healthcare I'm not aware of? It's really been accidental and it almost didn't happen. When I joined Microsoft in 2010, it was to join this great organization called Microsoft Research. I was very proud to be part of Microsoft Research,
Starting point is 00:29:42 and I rose through the ranks to the point where I was the leader of Microsoft Research worldwide. And then in 2016, Sachin Nadella and the CTO at that time, Harry Shum, reassigned me, took me out of research, and asked me to take on this 100 person team, a Skunkworks team to rethink Microsoft's approach to healthcare and healthcare technology.
Starting point is 00:30:10 I was devastated by that reassignment. I thought in fact I was being punished for some reason and actually contemplated quitting. You know, not only did I not have any background in healthcare or medicine, but Microsoft, you Microsoft, if you go to any healthcare organization, any clinic around the world, you will see Microsoft products there. We sell to literally
Starting point is 00:30:31 every single healthcare organization on the planet. I think one of our smallest accounts is a one nurse clinic in Nairobi, Kenya, and then all the way to giants like United Health Group or Kaiser Permanente and everything in between. And so that meant there had to be maybe a dozen powerful corporate vice presidents all throughout Microsoft, all doing their own things in healthcare. And so I also had to think, who's going to listen to Peter Lee on anything?
Starting point is 00:30:59 And so that's how it started. And you had to kind of think, what would we do? One question, Satya was worried that we weren't thinking enough about the cloud and AI in health care. So the first thing was to try to think, well, can the cloud be used to store health care records? And we learned early on that no, that there were compliance falls. There are certain data standards that we weren't supporting. And not only could our cloud not do it, but our competitors clouds, like Google
Starting point is 00:31:36 and Amazon couldn't do it either. And so that at least got us started with something to do, something to fix. I mean, we made a lot of progress in that. We also started a second project in collaboration with a company called Nuance and a doctor at University of Pittsburgh Medical Center, Shiv Rao, and that project was called Empower MD. Because we learned that doctors were really suffering with
Starting point is 00:31:59 having to write clinical notes to enter into electronic health records of them after every conversation with a patient. We thought, well, we could use AI to listen to the conversation and then at least draft a clinical note automatically. And so that was a project called the PowerMD. And that got serious enough that we actually went ahead and decided to productize that and acquire nuance in the process. And for Shiv Rao, Dr. Shiv Rao at UPMC, the venture arm of UPMC, UPMC Enterprises,
Starting point is 00:32:31 agreed to provide seed funding for Shiv to spin off a company to do the same thing. And that's a company called The Bridge. And today, the top two products in that space are Microsoft's DAXPilot and a product called The Bridge. That's how we got started. In that process of about five years of working on that, I got up to speed a lot. I also became a founding board member for a new medical school, the Kaiser Permanente
Starting point is 00:33:03 School of Medicine. A new school of medicine has to come up with a curriculum. And so I was able to study the curriculum and at least in pre-clinical studies, I learned quite a bit, all the way to the point that I actually got elected to the National Academy of Medicine. And then finally, Kevin Scott, our CTO in 2020, hired me back into research. And so I thought, okay, I can separate from healthcare, get back to what I, my true love, which is fundamental research and computer science, and then the pandemic hit.
Starting point is 00:33:36 So then Microsoft decided, well, Peter, you're our healthcare technology guy, and all of our customers and stakeholders need help from Microsoft to cope with the pandemic. So you need to coordinate that. So it kept me in the health care. I thought that that project will only last a summer of 2020, but of course the pandemic ended up being much more
Starting point is 00:33:56 serious thing than that. And then when GPT 3.5 and GPT 4 came out, there was again the question, wait a minute, is this stuff good to use as a safe to use in healthcare and medicine and I was the logical person to help lead that. I guess you could say I've been trying hard not to be in healthcare but I keep getting pulled back in.
Starting point is 00:34:21 I'm not unhappy about that, but it's odd because none of it has been planned. Some of the best careers are not planned and I think that's the testament to that. But I'm glad I didn't leave Microsoft. That would have been a good thing. I think we are all glad you didn't leave. That would have been, if you felt you were being punished
Starting point is 00:34:37 and quit immediately, we would have definitely lost out. As we get ready to close. Well, you'll go and notice here because Satya Nadella is such a remarkable leader. But I think what people on the outside sometimes don't know is he sometimes does ask people to do very, very hard things and it really moves things forward. That is true. When you're called to serve though, you either step up or you don't and it's good that you stepped up.
Starting point is 00:35:00 I was going to say that as we get ready to close, here we are in the middle of March. What are the things that are coming out of Microsoft Research that we should learn about as we close this podcast? What are some things that have either been announced or are being announced here at the middle of March? Right. We've been doing so much, particularly in two areas. One is what we would call AI for science. So in the same way that we've discovered that generative AI architectures, you know, like transformers and diffusion based models seem to learn so effectively from our words and thoughts and actions. So, you know, you can take a big
Starting point is 00:35:40 corpus of human text output or word output, and it's amazing what is learned there, and similarly from pictures of what we do out in the world and videos. Those same architectures, what the world is learning, and is a subject of huge intensity in Microsoft research, also work for observations of natural phenomena, like atmospheric wind patterns or the dynamics of proteins and small molecules or the movement of electrons in electrolytic material structures. And that is really incredible because that
Starting point is 00:36:20 means that if we follow the same path of AI scale in those areas, we might be able to do things like predict severe weather events weeks in advance or design a drug molecule for known drug targets or identify new drug targets in pathogens or be able to design new materials for everything from, say, solid-state batteries to enzymes to make your vegan food taste better. That AI for science thing, I think, is something that was a big showcase effort with a whole bunch of new models that are now in
Starting point is 00:36:59 the Azure Foundry coming out of Microsoft Research, while being simultaneously published in the top scientific journals in material science, in chemical engineering, in physics, in climate science, and so on. The one big difference, the one thing that holds us back in all of that is access to training data. There's no internet of molecular dynamics simulations.
Starting point is 00:37:22 The question is, where do you get the training data? We have the compute infrastructure, but we need the training data. And on that, the second thing that's emerging is quantum computing. And I think in 2025, we are going to see the first practical scalable quantum machines and the very first application that at least I
Starting point is 00:37:44 and some of our colleagues in Microsoft want to see is to run classical precise simulations of the natural phenomena to generate large amounts of perfectly labeled training data. And if we can do that, then we can literally have the GPT-4 or GPT-5 of proteins, of materials, of weather patterns. And I think that'll be pretty stunning. Yeah. And then that research to practicality bridge that like it doesn't just become theoretical at that point.
Starting point is 00:38:16 Like I love that you're putting out papers and simultaneously releasing a model on the foundry because I go back, call back to the whole beginning of this conversation. I've been told that my diabetes will be cured in five years, every year for the last 30 years. You know, show me the money, show me the practical thing that's going to prevent someone from losing their home in a tornado or prevent someone from dying of glioblastoma. You know, those kind of practical things. You're saying good things are coming from AI and from these models.
Starting point is 00:38:46 The way I've tried to explain it is that across many, many scientific domains, we have achieved GPT-2 level of capability. The only thing that's really preventing us to get to GPT-3 is access to adequate training corpus. The minute that we're able to solve those issues, we'll get to GPT-3 class capability and beyond. Then GPT-3 for us at Microsoft,
Starting point is 00:39:13 for you and me, Scott, has been important because GPT-3 and large language models was the first stage where we could try to make a product out of this. That, of course, was the first GitHub copilot. Yeah. That's the beginning of the hockey stick. When it starts to curve, then things start happening. Well, thank you so much, Dr. Peter Lee, for chatting with us today.
Starting point is 00:39:33 We really appreciate it. Well, Scott, thanks for having me here. It was really fun to chat with you. We have been chatting with Dr. Peter Lee, the President of Microsoft Research, and this has been another episode of Hansel Minutes in association with the ACM ByteCast. And we'll see you again next week.
Starting point is 00:39:48 ACM ByteCast is a production of the Association for Computing Machinery's Practitioner Board. To learn more about ACM and its activities, visit ACM.org. For more information about this and other episodes, please do visit our website at learning.acm.org slash bytecast. That's B-Y-T-E-C-A-S-T. Learning dot A-C-M dot org slash bytecast.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.