ACM ByteCast - Peter Lee - Episode 68
Episode Date: April 17, 2025In this episode of ACM ByteCast, our special guest host Scott Hanselman (of The Hanselminutes Podcast) welcomes ACM Fellow Peter Lee, President of Microsoft Research. As leader of Microsoft Research, ...Peter incubates new research-powered products and lines of business in areas such as AI, computing foundations, health, and life sciences. Before Microsoft, he established a new technology office that created operational capabilities in ML, data science, and computational social science at DARPA, and before that he was head of the CS department at CMU. Peter served on President Obama’s Commission on Enhancing National Cybersecurity and has testified before both the US House Science and Technology Committee and the US Senate Commerce Committee. He coauthored the bestselling book The AI Revolution in Medicine: GPT-4 and Beyond. In 2024, he was named by Time magazine as one of the 100 most influential people in health and life sciences. In the interview, Peter reflects on his 40+ years in computer science, from working on PDP-11s and Commodore Amigas to modern AI advancements. He highlights how modern technologies, built on decades of research, have become indispensable. He also talks about his healthcare journey, including work that earned him election to the National Academy of Medicine, and the potential (and limitations) of AI in medicine. Peter and Scott touch on the impact of LLMs, the lack of ethics education in traditional CS curricula, the challenges posed by growing AI complexity. Peter also highlights some important Microsoft Research work in AI for Science and Quantum Computing.
Transcript
Discussion (0)
This is ACM Bytecast, a podcast series from the Association for Computing Machinery, the
world's largest education and scientific computing society.
We talk to researchers, practitioners, and innovators who are at the intersection of
computing research and practice.
They share their experiences, the lessons they've learned, and their own visions for
the future of computing.
I'm your host today, Scott Hanselman.
Hi, I'm Scott Hanselman. Hi, I'm Scott Hanselman.
This is another episode of Hansel Minutes,
an association with the ACM Bytcast.
Today, I'm chatting with Dr. Peter Lee,
President of Microsoft Research.
You've got a resume as long as my arm,
and it is an absolute joy to chat with you today, sir.
It's a joy for me to be here. I really appreciate it.
I want to go in a little bit of a controversial direction because I want to start out with the fact that I had a birthday last week.
And in this year, if I understand correctly, you're going to turn 65. Is that correct, sir?
Yes, which I can't believe.
Yeah. Yeah. I turned 51 last week and I have been reflecting on the space that I've been in, in the last 32 years of software experience and you have more.
What does that feel like?
Because I've got, if you can see behind me,
I've got a PDP-11 that I built from a Raspberry Pi,
and I've been learning about that stuff.
I've got my Commodore 64 that I had as a child.
I'm really reflecting on how far this has come.
With the experience and the space that you have
occupied in computer science over the last 40 plus years,
what does that feel like as we sit here on this AI moment?
Yeah. First off, since you mentioned
PDB-11s and Commodore Amigas,
there's a lot of fondness in my heart for those.
I actually, my first real paying job was assistant administrator for a PDP 11.
And so I remember learning how to wire wrap the core memories there.
And then for Commodore, Amiga, Amiga NUT 64, took time away from my graduate studies at Michigan to be part
of a startup.
In that startup, one thing we were trying to do was put productivity software, like
word processing and spreadsheets and paint programs, onto Apple IIe and Commodore Amiga.
So if you ever used Amigarrite, it has MyCode.
Really?
And that company didn't really succeed, but we had to do things like make our own mouse,
because of course those computers didn't have mice at that time, but you needed a mouse to use those sorts of modern productivity programs.
And so really a lot of fondness for those. I reflected about this and Scott,
I assume you have a mobile phone in your pocket.
Of course.
The reason I think you have that is you look pretty
comfortable and relaxed because I think nowadays,
if you forgot it at home or left it in a hotel room,
you feel bad, you feel naked and
vulnerable like you can't really function properly.
And from a research perspective, I count six major triumphs of computer science
research that are in that mobile phone.
There's VLSI design that emerged in the 1970s.
There's a Linux or Unix inspired operating system kernel.
There is a software defined mobile wireless radio and so on. And these things were hardcore academic research
in computer science departments,
great ones around the world, in great laboratories,
you know, like Bell Labs and others.
They got published and then they made it into a suite
of technologies that you literally feel like you can't
live without all day every day. And you know it's kind of amazing when you
reflect on things like that. And so when I look at for example what's happening
today with LLMs in generative AI, the question in my mind isn't one about AGI or not. It's about when and if say large language model technology will become yet
another example of a technology that you can't bear to be without at any point
of your waking hours.
And I think it might happen, but you know, just, you know, as I've get closer to retirement at some point and just reflect
on everything that's happened, it's really incredible how much things have evolved to
the point where we really literally depend on them all the time for our mental health
and sanity.
Yeah.
That idea of not being able to live without it is such an interesting way of phrasing
it.
I have a very personal relationship with my phone
because it also runs my artificial pancreas.
And I was in Johannesburg last week for two weeks
and I bought a backup phone because even though
I can drive stick shift and just take needles
and take blood sugar, you know, with my finger,
like I could do that.
And I did that for 20 years.
Why is it now that I feel like I cannot survive without it?
I have to have the phone within Bluetooth range, you know, near me 24 hours a day.
And if I lost the phone or if it was dropped or stolen, I would immediately need to hook
another phone up because I want my continuous glucose meter.
And my wife, who's not diabetic but is a nurse, said, well, you could just take needles and stick your finger
like you did back in the 90s.
And I was like, that's unthinkable.
How dare you?
That's actually something called the flip.
And you see the flip most vividly
in the technology adoption history
in healthcare and medicine.
And my favorite example of that is ultrasound.
So, you know, ultrasound had been invented quite a long time ago, but it wasn't until the late 1950s that technologists proposed that ultrasound might be a good
diagnostic tool in the practice of medicine.
And so, you know, if you're a pregnant woman and you go to see your gynecologist
or obstetrician, you will see ultrasound used to
examine your unborn fetus
and the reproductive system of the mother.
But in the late 1950s,
when this was first proposed, it scared people.
Is this going to damage my baby?
Is it going to damage my ability to reproduce?
Are doctors trained to read
these grainy images and all sorts of problems.
And so it took about a decade before ultrasound became the standard of care.
And so now, what do I mean by the flip?
Well today, if you are a pregnant woman and you go to your obstetrician and your obstetrician
says, ah, I don't believe in that fancy ultrasound technology.
I'm going to put my hands on your belly and use manual palpation."
Not only would you be horrified at that, but you might even report that as malpractice.
And so expectations, not only in the standard of care, but in patient demands and expectations,
consistently flip in the history of technology adoption medicine.
And your example is another example of how things flip.
You flip from being able to drive manually
and monitor your own blood sugar levels
to a feeling that why would you do that?
It's so primitive, it's so inexact, so dangerous,
unconscionable, and in fact irresponsible to do it that way.
I'm curious though that I find myself feeling like old man who shakes fist at cloud because
I feel like the young people are saying it's going to change the world and if you're okay,
boomer, if you're against it, then if you're not with us, you're against us.
But I remember when the TI-83 and the TI-81 calculators came out and the math teachers
were like, no one's going to be able to do math in their head anymore
because you won't always have a calculator young man,
and here we are with a pocket supercomputer.
I am finding myself questioning my own opinions about tech based on my age and
my generation and I'm really trying to be introspective.
I get that it's going to change the world,
but I also think that cognitively,
we don't know how these's going to change the world, but I also think that cognitively,
we don't know how these are going to change our brains.
And I don't know if having a pocket supercomputer has changed our brains for the positive.
Yeah.
You know, I'm always a little bit more circumspect.
It's easy as two people who are tech executives, you know, to look at all the historical examples
of, you know, so the adoption of advanced technologies that are really enabling,
there is a consistent pattern.
And so you could always argue, as we in tech do,
that the right side of history is that these things advance
and the world gets better and better.
And I think that's generally true, but I think it's also true,
and it's worth some reflection
that we lose some things along the way. And that isn't to say that the world isn't getting better
and life isn't getting better because I think it is by any measure. But I think it's wrong to deny
that we're losing some things and some skills and some abilities that maybe were important. One thing I would say about,
especially in this new AI era is,
I do see that the world's leading thinkers are
actually earnestly trying to be thoughtful about this.
That there is a big debate.
In fact, it's a debate of leading thinkers that I haven't seen
since the human genome was mapped.
When we finished mapping human genome
as a scientific community,
it sparked a huge debate about what would this mean?
And would genetic engineering be a good or a bad thing?
And the fact that there were so many leading thinkers
and the whole academic and research area of bioethics
really just mushroomed into hugely important kinds of
thought leadership and research.
I think that's helped us a lot to gain as much of the goodness out of our understanding
of genetics while mitigating the potential downstream harms and risks.
And I see the same intensity of debate going on with AI
as people are trying to grapple with what this means.
I think that that's a good thing and it actually makes it
easier for me to be optimistic about what's happening right now.
Yeah. As someone who is and was
a professor of computer science for many, many years,
did you plug ethics in and
did you have people thinking about the tech?
I mean, sometimes it's fun to just do techie stuff for techie reasons, but I've always come back to,
are we making someone's life better by doing this or are we making someone's life demonstrably worse?
But a lot of people, when I do these informal polls, I say, all right, who here has ever taken
a computer science ethics class? The hands are not going up, a majority or plurality.
It feels like we're not teaching computer science ethics
and it feels like it's even more important
in a post-GPT-4 world.
Yeah, I have to admit, earlier in my career,
it really wasn't a serious thought.
In fact, if anything, I tried hard to adopt
the Silicon Valley stereotype ethos of tech is good and more tech is better.
And just the belief that technology could be the answer to all of our problems and
that we can solve all problems through technology.
I do think we've gotten to such a more enlightened state about this.
I think a lot of that has come through lessons learned
the hard way. We've seen so clearly that tech isn't different than any other technology
in being dual edged. And in fact, it's more interesting and has greater positive and negative
potential because information technology can be democratized. It can be given to literally
everyone.
You know, unlike genetic engineering where you have to have a multi-million dollar wet lab to do anything here,
literally every person on the planet could gain access and harness this for both good and evil.
And so I think ethics has become a much bigger deal. Whereas I think earlier in my career, personally,
I think it was an age of innocence, where you only saw the goodness in things.
Yeah. I really appreciate the democratization of it.
Open source itself,
the pancreas that I use is
a toolkit that one could build from scratch,
and I did, and that anyone could go and do that.
Now we're seeing open weights and open models.
What's your thinking about frontier models that are
closed versus frontier models that are more open
about where they came from, what they were trained on, and the corpus that built them up?
Yeah, I think that this is also evolving really quickly. The thing that's so interesting to me is
the cost of doing something first versus the cost of doing something
versus the cost of doing something tenth. And so if you are OpenAI or one of OpenAI's frontier competitors,
you're trying to get to certain levels of intelligent capability first.
And there you can see a pretty methodical investment strategy.
To get to GPT-2 maybe requires on the order of $5
to $10 million of compute cost to train a model at that level.
And GPT-2 allows you to see perfect loss curves.
So that gives you confidence that further scaling is going to get you somewhere.
And then to get to GPT-3 level and to do that first, pretty consistently in whatever domain
you're trying to train your A
models consistently requires about 10x the compute cost. So now you're talking 50 to 100 million
dollars and then to get to GPT-4 level you know whether it's in language or molecular dynamics
or you know weather modeling whatever seems to take yet another 10x leap if you want to get there first.
So now, you know, you're starting to get close to a billion dollars and so on.
The thing that's so interesting is, like in all things that have to do with technological invention,
once smart people see that something is possible, then it becomes so much easier for the second and
third and fourth and tenth people to do it.
And what we're seeing in the industry is that not being first means that you can also,
of course, you don't benefit from being first, but you benefit from being able to be smarter and a hell of a lot cheaper to get to the same state. Not that getting to a GPT-4 class base model is cheap.
It's still quite expensive, but it's a lot cheaper
than it took for the very first innovators to get there.
This is the pattern of tech that you and I have lived through
our entire careers.
So we shouldn't be surprised by this,
but just to see it playing out with just this stunning speed
is just incredible.
And business-wise, the question is,
there's always been a big premium in our business in being first.
Now the question is, given the rapid pace of change and evolution,
is there still the same amount of value in being first as there has always been?
The idea that we shouldn't be surprised,
but we consistently are surprised.
I remember the first time that I saw, you know, a
chat bot for lack of better words.
And now like I'm looking at my phone, I've got an
app called MLC chat and I've got five three, five
mini it's a, a four bit quantized F 16, but I'm
having a conversation in airplane mode on a, you
know, on a 15 tops iPhone.
That seems insane to me and I get it and I understand it.
But we're at a point now where the full stack, you
know, you always hear about the full stack engineer
is so deep that I find myself going back to building
things from original parts, just so I can remind
myself, like I've got a, an Apple one that I'm
building from 7,400 series chips,
just to kind of get my hands back in the silicon
because the stack is so high,
I'm trying to get my hands dirty
so that I can emotionally like accept
that there's an SLM on my phone
that works in airplane mode.
You know, by the way,
I'll take one of those Apple Ones
if you make two of them.
Yeah, yeah, those are great.
That's a company called Smarty Kit.
I'll send that to you.
You know, there's something that you're saying there though that is another reflection just looking back at my career.
And I think you and I, when we started in all of this,
we were able to wrap our heads around and know the intricate details of the entire stack end to end.
From the silicon, or in my case, the wire wraps,
all the way to all the code, every line of code
in the applications.
And things are just so much more complicated now.
And when we think about what AI is going to enable,
I think AI is very quickly going to enable us to construct
systems that are so complex that they will really completely defy any ability for humans to
comprehend them fully. In research, where I think this might happen first, Right now, there's a vibrant set of researchers that are using generative
AI to write mathematical proofs. And when you ask an AI system to write a mathematical proof,
you generally ask it to write it in a proof language, a popular one is called Lean. There
are others as well. The interesting thing about a proof language is that they're set up so that you can use a simple
type checker, just like you would have in any programming language. And these proof languages
are set up so that if they type check, you know for certain that the proof is valid. And so I foresee
in much less than five years that we'll have an AI system generate a proof of some mathematical
theorem. We'll be able to type check that proof to know that that proof is absolutely valid and
correct. But that proof itself might defy any ability for any human being, even the world's
smartest human beings, to understand it. And that's sort of like a point example of AI getting us to a place where we're able to construct things, build things,
that work. And we know we can see them work, we can validate that they work,
but we won't know how or why. And I think it's going to be, when people ask me about AI or superintelligence,
that will be the first mark of it for me.
And, you know, your roots and my roots, we still want to, you know, understand everything.
And we do things to keep that fresh for us. But I think it's going to get hard.
ACM Bytecast is available on Apple Podcasts, Google Podcasts,
Podbean, Spotify, Stitcher, and TuneIn.
If you're enjoying this episode,
please do subscribe and leave us
a review on your favorite platform.
That's a really interesting point.
I think you're right that I and perhaps people of our generation plus or
minus a few years are still unwilling to let go of the fact that we took a rock,
we flattened it, we infused it with lightning,
and now it talks to us and I want to
understand the lightning and the squishy rock part.
But my children, my 19-year-old,
is perfectly willing to just
accept that the magic black box is doing a thing.
When I have to have conversations with him about maybe,
let's not anthropomorphize the AI,
let's talk about what's really happening.
He's like, it helped me with my homework and I'm cool with that.
I was a little bit taken by your book,
The AI Revolution in Medicine that was pre-GPT-4 general availability.
You were poking at the model in ways where you're like,
you were anthropomorphizing it, but you were also trying to understand it.
What would you change about how you interacted with the model
now that you know more, since this very good book has been published?
Yeah, you know, that book, you know, we wrote it over the Christmas holidays in 2022,
you know, while GPT-4 was still a secret project.
And we organized the ourselves so that the book would get published at the same month that GPT-4
was released to the whole world. And so that was a time when, you know, we were just so amazed
and baffled. You know, I talk about this as the nine stages of grief. In fact, I think at some point in the book, I talk about the nine stages of AI grief.
When GPT-4 was first exposed to me by the folks at OpenAI, I was intensely skeptical
because they were just claiming that this thing could do things that I just thought
were not possible.
And then you pass from that stage of skepticism to the stage of frustration,
because I felt like I was seeing my colleagues in Microsoft research
falling, getting duped by this thing.
And then you start to feel worried because I detected that,
wow, Microsoft might actually make a big bet on this thing.
But then you get hands-on and you start to encounter things that are just amazing.
And I remember feeling the joy that, wow, this thing is, I never thought I would live long enough
to see such a technology little and have my hands on it. And then you get into a period of
intensity. So there are these stages that you go through, but in those early stages of joy and euphoria and you lose sleep, it is those aspects that
feel empathetic that draw you into anthropomorphism that end up being so interesting.
And in the field of medicine, this has been observed over and over again.
In fact, one month after we published our book, UC San Diego and Stanford jointly published a paper
in a medical journal where they used GPT-4 to respond to emails from patients.
They compared those to the emails that doctors, human doctors, would write.
Then they had a blind test and they had patients and doctors grade the quality, correctness
of these responses to patient
queries.
Not only was the AI equally accurate, but by a factor of nine to one, the AI-generated
notes were judged by patients to be more empathetic.
And of course, you know, it seems crazy to say a machine can be empathetic. What it really means, I think,
is that a frazzled doctor can't take
the time to write more than two or three sentences
and just get to the point and then get onto the next email.
Whereas the AI can write a couple of paragraphs and might
remember that during the encounter they were talking about
going to a Seahawks game and other stuff like that and
put in those nice personal touches.
And so there's something there that is both interesting and disturbing,
but also seems to really touch people in a very meaningful and very practical way.
And I think we're still, just as a we're trying to come to grips with this.
Yeah. I'm struck that one could theoretically say that it can be
infinitely empathetic and infinitely patient given
appropriate prompting and an appropriate good attitude on the part of the controlling human.
I keep coming back to empathy.
We need more empathy in the world right now,
we need more empathy in tech.
When I use Copilot on GitHub Copilot, I don't ask it how to do my homework.
I basically use it as an enthusiastic pair programming partner and I find it to be infinitely
patient.
It never judges me, it's never mean, it's never unkind.
So then it has theoretically infinite empathy if I talk to it right.
I'm struck by the epilogue in your book.
You say, it just can't be that
next word prediction could be intelligence or can it?
Am I just a statistical model of what's
the most likely next word from Hansel and to say?
Is that the animus of me?
I think, yeah, at the time we wrote that book, you know, this was, I was
really baffled by this, but I think my understanding and acceptance of what's
going on has evolved a lot more.
It's true that the fundamental pre-training of these large language
models is to predict the next word in the conversation.
And so to that extent, you could say
that these language models aren't trained
to do anything useful except that.
But here's an example that I like to give.
Let's take the sentence, and the murderer is blank.
Okay, so now if you want to, you know,
pick with the highest quality, the word that would fill in the blank, well
that sentence, and the murderer is blank, is in the context of let's say a whole Agatha
Christie murder mystery novel.
And if you were to just approach this purely as a statistics question, well there are thousands
of murder mystery novels and short stories, and you could try to make a statistical pattern on what are the most likely names of murderers, and you wouldn't
get any good answer there.
Instead, in order to really optimize the quality of the fill-in-the-blank capability, you somehow
have to be able to do some deductive reasoning.
You have to have an understanding of the psychology of humans in different situations,
what motivates them, how they react under certain kinds of questioning, and all of those things.
And so the way to understand what's going on when we try to optimize,
fill in the blank or next word prediction, is as a very, very large,
astronomically large stochastic process that has a chance
of accidentally discovering neural circuitry that implements some aspects of those reasoning
functions.
And the fact that that can happen at all, even by accident, is amazing. But we're operating a scale where, indeed, it is actually happening.
And so it's not that it's next word prediction that is causing us to appear
to be thinking, but the process of highly optimizing an ability to do very,
very good next word prediction is giving us a chance to really discover and solidify these
bits of neural circuitry to do things.
Yeah.
One analogy that I've used to explain to some young people, and I don't know if it's a good
one, is that we are, as humans, limited by the size of our stack for our recency.
And then the older you get, you get this larger and larger and larger heap that you can pull
from. And then you're constantly pulling things out of the heap and into the stack.
And then there's certain presenters, there's certain thinkers.
Jamel Bouie is one I think of who's an opinion person for the New York Times,
who seems to have this huge corpus of information that he is pulling upon all the books that he's ever read.
He's had an amazing vocabulary.
And I admire him and his ability to page in and out
these pieces of wisdom while I struggle to root
around in the totality of my existence.
I think that AI will feel like AGI when its stack,
when its context window is
beyond those of even the smartest person,
and that's going to be the thing, that it has just such a large context window, it's bigger even the smartest person. That's going to be the thing,
that it has just such a large context window,
it's bigger than a human lifetime.
Yeah. One thing that up till now,
for the most part, large language models and
transformers specifically were pretty imperfect
in memorization of things.
There's a massive compression of the
training corpus that goes on when you train it into a transformer. And so, you
know, I was always trying to explain that to doctors as they're trying to come to
grips with generative AI because, you know, unlike a normal computer, in fact,
one popular application, maybe the most popular
application that doctors use online is something called UpToDate. It's essentially a search engine
for very highly curated medical knowledge. And so in UpToDate, you ask a question and you get a
medically precise answer. And we've trained ourselves to use that just like we use web search.
You make a query and you expect to get a set of answers that are pretty precise.
But the transformer doesn't have that capability.
It has very, very imperfect memory.
Of course, that's now evolving because there's no fundamental reason why
computer-based system
couldn't actually have perfect memory recall. And so I agree with you. I think we're going to get to
a point where these AI systems really are going to be benefiting from the fundamental capabilities
of perfect memory recall that we've always assumed machines would have, while also making all of
these associations and engaging in this reasoning.
Why is it, and forgive my ignorance if this is
a question I should know the answer to,
but when did your deep interest in
the medical aspect of technology,
like in everything that you do,
and certainly recently of late,
you always want to make
technology to make people's lives better,
and it seems to come back to healthcare.
Do you have a background in healthcare I'm not aware of?
It's really been accidental and it almost didn't happen.
When I joined Microsoft in 2010,
it was to join this great organization called Microsoft Research.
I was very proud to be part of Microsoft Research,
and I rose through the ranks to the point where I was
the leader of Microsoft Research worldwide.
And then in 2016, Sachin Nadella and the CTO at that time,
Harry Shum, reassigned me, took me out of research,
and asked me to take on this 100 person team,
a Skunkworks team
to rethink Microsoft's approach to healthcare
and healthcare technology.
I was devastated by that reassignment.
I thought in fact I was being punished for some reason
and actually contemplated quitting.
You know, not only did I not have any background
in healthcare or medicine, but Microsoft, you Microsoft, if you go to any healthcare organization,
any clinic around the world,
you will see Microsoft products there.
We sell to literally
every single healthcare organization on the planet.
I think one of our smallest accounts is
a one nurse clinic in Nairobi, Kenya,
and then all the way to giants like United Health Group
or Kaiser Permanente and everything in between.
And so that meant there had to be maybe a dozen powerful corporate vice presidents
all throughout Microsoft, all doing their own things in healthcare.
And so I also had to think, who's going to listen to Peter Lee on anything?
And so that's how it started. And you had to kind of think, what would we do?
One question, Satya was worried that we
weren't thinking enough about the cloud and AI in health care.
So the first thing was to try to think, well,
can the cloud be used to store health care records?
And we learned early on that no, that there were compliance falls.
There are certain data standards that we weren't supporting.
And not only could our cloud not do it, but our competitors clouds, like Google
and Amazon couldn't do it either.
And so that at least got us started with something to do, something to fix.
I mean, we made a lot of progress in that.
We also started a second project in collaboration with
a company called Nuance and
a doctor at University of Pittsburgh Medical Center,
Shiv Rao, and that project was called Empower MD.
Because we learned that doctors were really suffering with
having to write clinical notes to enter into
electronic health records of them
after every conversation with a patient. We thought, well, we could use AI to listen
to the conversation and then at least draft a clinical note automatically. And
so that was a project called the PowerMD. And that got serious enough that we
actually went ahead and decided to productize that and acquire nuance in the process.
And for Shiv Rao, Dr. Shiv Rao at UPMC,
the venture arm of UPMC, UPMC Enterprises,
agreed to provide seed funding for Shiv to spin off a company to do the same thing.
And that's a company called The Bridge.
And today, the top two products in that space are
Microsoft's DAXPilot and a product
called The Bridge.
That's how we got started.
In that process of about five years of working on that, I got up to speed a lot.
I also became a founding board member for a new medical school, the Kaiser Permanente
School of Medicine. A new school of medicine has to come up with a curriculum.
And so I was able to study the curriculum and at least in pre-clinical studies,
I learned quite a bit, all the way to the point that I actually got elected to the National Academy of Medicine.
And then finally, Kevin Scott, our CTO in 2020, hired me back into research.
And so I thought, okay, I can separate from healthcare,
get back to what I, my true love,
which is fundamental research and computer science,
and then the pandemic hit.
So then Microsoft decided, well, Peter,
you're our healthcare technology guy,
and all of our customers and stakeholders
need help from Microsoft to cope with the pandemic.
So you need to coordinate that.
So it kept me in the health care.
I thought that that project will only last a summer of 2020,
but of course the pandemic ended up being much more
serious thing than that.
And then when GPT 3.5 and GPT 4 came out,
there was again the question,
wait a minute, is this stuff good to use as a safe to use
in healthcare and medicine and I was
the logical person to help lead that.
I guess you could say I've been trying hard not to be in
healthcare but I keep getting pulled back in.
I'm not unhappy about that,
but it's odd because none of it has been planned.
Some of the best careers are not planned
and I think that's the testament to that.
But I'm glad I didn't leave Microsoft.
That would have been a good thing.
I think we are all glad you didn't leave.
That would have been, if you felt you were being punished
and quit immediately, we would have definitely lost out.
As we get ready to close.
Well, you'll go and notice here
because Satya Nadella is such a remarkable leader.
But I think what people on the outside sometimes don't know is he sometimes does ask people to
do very, very hard things and it really moves things forward.
That is true. When you're called to serve though,
you either step up or you don't and it's good that you stepped up.
I was going to say that as we get ready to close,
here we are in the middle of March. What are the things that are coming out of Microsoft Research
that we should learn about as we close this podcast? What are some things that have either
been announced or are being announced here at the middle of March?
Right. We've been doing so much, particularly in two areas. One is what we would call AI
for science. So in the same way that we've discovered that
generative AI architectures, you know, like transformers and diffusion based models seem
to learn so effectively from our words and thoughts and actions. So, you know, you can take a big
corpus of human text output or word output, and it's amazing what is learned there,
and similarly from pictures of what we do out in the world and videos.
Those same architectures, what the world is learning,
and is a subject of huge intensity in Microsoft research,
also work for observations of natural phenomena,
like atmospheric wind patterns or the
dynamics of proteins and small molecules or the movement of electrons in
electrolytic material structures. And that is really incredible because that
means that if we follow the same path of AI scale in those areas, we might be able to do things like
predict severe weather events weeks in advance or design a drug molecule for known drug targets or
identify new drug targets in pathogens or be able to design new materials for everything from,
say, solid-state batteries to
enzymes to make your vegan food taste better.
That AI for science thing, I think,
is something that was a big showcase effort
with a whole bunch of new models that are now in
the Azure Foundry coming out of Microsoft Research,
while being simultaneously published in
the top scientific journals in material science,
in chemical engineering, in physics,
in climate science, and so on.
The one big difference,
the one thing that holds us back in all of that is access to training data.
There's no internet of molecular dynamics simulations.
The question is, where do you get the training data?
We have the compute infrastructure,
but we need the training data.
And on that, the second thing that's emerging
is quantum computing.
And I think in 2025, we are going to see
the first practical scalable quantum machines
and the very first application that at least I
and some of our colleagues in Microsoft want to see
is to run classical precise simulations of the natural phenomena to generate large amounts of
perfectly labeled training data. And if we can do that, then we can literally have the GPT-4 or GPT-5
of proteins, of materials, of weather patterns.
And I think that'll be pretty stunning.
Yeah.
And then that research to practicality bridge that like it doesn't just become theoretical
at that point.
Like I love that you're putting out papers and simultaneously releasing a model on the
foundry because I go back, call back to the whole beginning of this conversation.
I've been told that my diabetes will be cured in five years, every year for the last 30
years.
You know, show me the money, show me the practical thing that's going to prevent someone from
losing their home in a tornado or prevent someone from dying of glioblastoma.
You know, those kind of practical things.
You're saying good things are coming from AI and from these models.
The way I've tried to explain it is that across many,
many scientific domains,
we have achieved GPT-2 level of capability.
The only thing that's really preventing us to get to
GPT-3 is access to adequate training corpus.
The minute that we're able to solve those issues,
we'll get to GPT-3 class capability and beyond.
Then GPT-3 for us at Microsoft,
for you and me, Scott, has been important because
GPT-3 and large language models was
the first stage where we could try to make a product out of this.
That, of course, was the first GitHub copilot.
Yeah. That's the beginning of the hockey stick.
When it starts to curve, then things start happening.
Well, thank you so much, Dr. Peter Lee,
for chatting with us today.
We really appreciate it.
Well, Scott, thanks for having me here.
It was really fun to chat with you.
We have been chatting with Dr. Peter Lee,
the President of Microsoft Research,
and this has been another episode of
Hansel Minutes in association with the ACM ByteCast.
And we'll see you again next week.
ACM ByteCast is a production of the Association for
Computing Machinery's Practitioner Board.
To learn more about ACM and its activities, visit ACM.org.
For more information about this and other episodes, please do
visit our website at learning.acm.org
slash bytecast.
That's B-Y-T-E-C-A-S-T.
Learning dot A-C-M dot org slash bytecast.