Behind The Tech with Kevin Scott - Dr. Eric Horvitz: Chief Scientific Officer, Microsoft

Episode Date: May 28, 2020

Microsoft’s first Chief Scientific Officer, Technical Fellow and world-renowned AI researcher, Dr. Eric Horvitz, gives us an insider’s look into how the experts are approaching the fight against C...ovid-19. Learn how AI and bioscience are playing a critical role in gaining insights into the best way to approach this pandemic and prepare for future challenges.   Click here for full transcript of this episode. Listen to other Microsoft podcasts at aka.ms/microsoft/podcasts

Transcript
Discussion (0)
Starting point is 00:00:00 I think that we know so little even today and there's so much possible that I think our grandkids will live in a very different world where biology will have played a major role in almost everything that's being touched including the materials that we build systems from and so I'm very excited about the possibilities materials that we build systems from. And so I'm very excited about the possibilities. Hi, everyone. Welcome to Behind the Tech. I'm your host, Kevin Scott, Chief Technology Officer for Microsoft. In this podcast, we're going to get behind the tech. We'll talk with some of the people who made our modern tech world possible and understand what motivated them to create what they did.
Starting point is 00:00:48 So join me to maybe learn a little bit about the history of computing and get a few behind-the-scenes insights into what's happening today. Stick around. Hello, and welcome to Behind the Tech. I'm Christina Warren, Senior Cloud Advocate at Microsoft. And I'm Kevin Scott. It is great to see you, Kevin. Well, you know, virtually anyway.
Starting point is 00:01:14 So are you getting into the rhythm working from home yet? I think it has finally gotten a little bit easier. The big challenge for us in the early days of the COVID shelter in place was having both of our kids home. So I've got a nine-year-old and and also doing our own IT support for our kids, which was not really what we had signed up for. Right, right. So how has your work changed? I know you traveled a lot. So that's obviously not happening anymore. So how's your team working?
Starting point is 00:02:07 Yeah, obviously I'm not traveling anymore, right? That's for sure. Yeah, you know, it's been interesting. Our team was mostly remote first anyway, but this is a different sort of remote experience. So it's been, I think, good in the sense that we already had experience working together from different places, but it's still different and we're all, I still think, adjusting. But one of the interesting things with Microsoft is we're focusing on trying to see how we can do advocacy and events and the things that I used to travel for online and potentially reach even more people.
Starting point is 00:02:41 So that's pretty cool. And as our listeners may know, Microsoft Research is actually playing a significant role in bringing our technical resources to the table. And so today we're going to hear from a Microsoft colleague who is leading the charge. Yeah, we're really excited to have Dr. Eric Horvitz on the show today. Eric's one of the most highly regarded AI researchers in the field with contributions in spam machine learning, perception, natural language understanding, and decision making. His efforts to understand the influences of AI on people in society, including ethics, law, and safety, really are paving the way for responsible AI practices. It's such important work. So let's hear what Eric's been up to.
Starting point is 00:03:38 Our guest today is Dr. Eric Horvitz. Eric is a Microsoft Technical Fellow and the company's first Chief Scientific Officer. As Microsoft's Chief Scientist, Eric provides leadership and expertise on a broad range of scientific and technical areas, from AI to biology and medicine, to a whole host of issues that lie at the intersection of technology, people, and society. Eric earned a PhD in AI from Stanford and is one of the field's leading innovators and luminaries. Eric also has the rare distinction of having earned his MD, also from Stanford, which gives him a unique view and understanding of the many connections between AI, biology, and biomedicine. I'm thrilled to have Eric with us today. Welcome, Eric.
Starting point is 00:04:13 Thanks, Kevin. It's great to be here. Yeah, so I'd love to start, as we always do, by understanding how it is you first got interested in science and technology. Presumably that was when you were a kid. So can you tell us a little bit about that? Yeah, I just know that I've always been sort of inspired to understand things, and I didn't distinguish between human creations, artifacts, and stuff I would see in the world. So I was confused and intrigued and interested in living things in space and time. I remember even being very, very young, asking my first grade teacher if I can know more about time. And she ended up bringing me to the library at Birch Elementary School and showing me a book about clocks.
Starting point is 00:05:03 And I said, no, I don't really mean clocks. I mean time. And I was also intrigued by light. I had this really beautiful phosphorescent, phospholuminescent nightlight in the 60s. And beautiful green light would wash the room at night in this glow. And I was just curious, what the heck was light? So, I had these basic questions. I remember having a discourse with my father about, you know, I heard a lot about God and I was curious what God was made of. And I couldn't get a good answer from adults about that. And when it comes to machines and mechanism, I took apart a flashlight, I think it was like the summer after kindergarten or so, because I remember in first grade, I was already into this and talking to friends about this. But I realized that there
Starting point is 00:05:49 was a circuit there and I found some wire. And I think I impressed my family more than myself when I ran around the house with a battery and a wire with a light bulb lighting up with my finger, under my finger. And I think this was also around the time that, again, mid-60s when there was a lot of, you know, a lot of the cartoons we were watching back then had electronic robots and Astro Boy flying around, very helpful entities. And I was curious about electronic brains. I don't know where I got that idea, but I remember having a bag of parts on my way to my grandmother's house in the back of a station wagon. Maybe this is around second or third grade, but the peanut can wires, light bulbs, I thought I could assemble an electronic brain on the way to my grandmother's house in the back of a station wagon. And didn't get, you know, still looking on that today,
Starting point is 00:06:48 basically. That's really awesome. And were your parents scientists or technical engineers? My parents were both school teachers. My mother was a kindergarten teacher. And I remember being very proud of that in kindergarten. I would tell everybody at a time when the kindergarten teacher was like the person you most looked up to that, by the way, my mom was a kindergarten teacher too. That was considered awesome by my peers at the time. And my father was a high school teacher. He did science as well as history. So where, I mean, it sounds like you had a bunch of innate curiosity, which is awesome. And like one of the themes I think we see with a lot of people who chose careers in science and technology. But did you have any role models when you were a little kid or things that were in the popular media that were inspiring you? Or did this just really come out of, you know, from your perspective, nowhere?
Starting point is 00:07:53 Lots of books. My parents had a home library filled with lots of books. We had the Merrick Library, Merrick, Long Island, where I would spend lots of time. I got to know the science sections as well as the pet section of the library pretty intensively. And mostly books at the time, and friends, some of whom had aligned interests. It's hard to think of the idea of being in the first or second grade and having a third grade, having a scientific support team. But we sort of had peers that were interested as well. In third grade, I was elected to be the chairperson of the science club, I remember. And we had all sorts of projects involving wind speed and solar energy back in those days.
Starting point is 00:08:40 But I'm not sure where some of the interest came from. It was largely curiosity and books. And later in life, of course, I had some fabulous mentors. And, you know, we all think back to our various teachers in elementary school. You know, you start in kindergarten and go to sixth grade. Each teacher has a major influence on people. And, you know, I can remember sitting at this desk in sort of a, I thought it was kind of a militaristic setting. And I asked myself on the first day of first grade, is this what school
Starting point is 00:09:12 is going to be like? I have to sit at this desk like for like, like 12 years. And I, the way that the first grade went, I was really unimpressed. And I would have given it all up if it wasn't for, and I'll call out a name, Mrs. Frank, my second grade teacher, who like completely opened the world to me. Was open to science and interested in answering questions. You know, and then you jump forward to fifth grade, Mrs. O'Hara. And these people were just brilliant teachers. Mr. Wilmot in sixth grade, where he celebrated my interests, and we had science fairs, and I actually won the science fair that year. And you have a few teachers like that, that really are like large planets that spin you off with the gravitational field into new directions. Yeah. You know, I think that's something that we, as a society, systemically underappreciate is the role of these really incredible teachers and what a massive influence they have in your life.
Starting point is 00:10:15 It's amazing. Yeah, I'm quite certain that if Mrs. Frank wasn't there in second grade, I'd be doing something very different now in the world. Yeah. Well, and it's also really interesting. I think to a certain extent, all children have this innate curiosity. So it'll sort of be interesting to talk about this later when we are chatting about AI. But in a sense, humans are learning machines machines and we sort of come into this world and we, you know, we have an innate curiosity to understand, you know, what's going on around us. to focus on with my children is to do everything that I possibly can to encourage them to lean into and celebrate their own curiosity and to support it in all of the ways that you can. Because I have this very strong belief that curiosity is one of these pivotal things that helps you be successful in life. And even when you're not talking about technology or science, you're talking about your fellow human beings and trying to develop things like compassion. I believe compassion is rooted in
Starting point is 00:11:39 curiosity. It is like you wanting to know where another person is coming from or like what they're thinking about or how they're processing their world. And so like I just – it's unbelievably important. I believe this curiosity. in your life who really celebrated that curiosity rather than thinking it was this annoying thing that was distracting them from what else they were trying to do. Yeah, it's kind of interesting. It's almost like there is something innate and basic in humans. And I've heard biologists and anthropologists talk about what makes Homo sapiens different than some other even closely related primates. And some of it is this delayed maturation.
Starting point is 00:12:33 They talk about this idea that human beings are more kid-like their whole life than closely related uh a species uh kid like referring to puppy-like curiosity that just continues on um this idea of continuing to explore versus being locked in and anything we can do to promote i think which is very much a human uh probably makes us more human than we know deep unrelenting curiosity i think can go a long way for individuals and for society as a whole. I was thinking about years ago, thinking about how much pleasure I get. It's almost like raw pleasure with getting an answer. This tension combined with a little bit of awe and mystery of a question building
Starting point is 00:13:25 and the pressure around that and how when it gets resolved into a partial answer, the gradient that you're on and the kind of pleasure you get traveling through that terrain is so deep and great. It's like one of the deepest pleasures I know. These bursts of insight. And to have to just think that in some people that might be linked to pain, or I don't want to go there. And the fact that that could come from the nurturing that led to that kind of shift of the natural pleasures of learning and growing
Starting point is 00:13:59 to a painful, I don't want to go there because I don't want to learn something new, for whatever reasons of background. It is very sad. Yeah. Well, I know even myself, like, there is a certain degree of discomfort to being fully immersed in a problem. Because, like, I don't know about you, I tend to get obsessed with questions and trying to find their answers. I remember when I was in grad school, I would be working on proving a theorem. And, you know, with some of these things I would spend days on. And I, on multiple occasions, I would be so immersed in the problem that I was trying to solve that I would dream about it. And several times I dreamt the solution to a theorem I was trying to prove. And I would wake up and like, oh my, I got it now.
Starting point is 00:14:56 And I would go write everything down before I forgot. And I experience that sometimes is discomfort. So I like totally understand what you're saying about this. You know, like sometimes maybe people experience a little bit of fear and anxiety when they are approaching an unknown. Right. And if they have to get used to the notion or get familiar with the idea that there are pleasurable bumps along the way and a pop towards the end when you get near a solution.
Starting point is 00:15:34 For me, it's similar. Sometimes I'll have a problem. I remember from my dissertation work, actually, at Stanford, I was really, at a time, worrying about this tension between how do you do things formally with probability and decision theory when it's intractable? And when you needed this kind of reasoning to do some good work in high stakes decision problems. And just being at loggerheads with contradiction. I remember actually where I was at the moment. I was visiting my family and cleaning the garage.
Starting point is 00:16:03 I'm not sure how they got me to do that on this day. But actually, I've had this image of looking at this stuff scattered throughout the garage and in my mind, seeing an interesting solution coming to the fore finally, just out of the blue, that became the kernel of what I ended up working on and the solution to this tension. So sometimes you run these batch jobs, which are just tantalizing in there in the background and popping up when you're driving a car or when you're cleaning a garage. But you're online and a portion of your soul is really focused on getting to an answer continually. Yeah. And I think the other thing that I will say and then start talking about your trajectory a little bit more, but I think there is a very interesting thing about this whole phenomenon that you are describing where you've got the discomfort of the unknown and this sort of tension between the thrill of discovery and the frustration of navigating a problem that you can get better at over time if you practice. So, like, the more you do it, the more that you understand that you are
Starting point is 00:17:16 going to be able to get these little victories over the problem and, like, hopefully, like, be able to get to a good solution at the end of the day. And I think as you understand that, it makes you more not just willing but eager to go seek these problems out because it really does become this amazing experience and and like very rewarding and i should say that it's not all individual um as i'm thinking about the the visceral sensations we have as we think about a problem or or ask a question and then pursue an answer urgently or over time um there's the sense that i've had i remember being looking at the stars one night as a young kid, maybe a little bit more into middle school and feeling anxiety about hanging in three space
Starting point is 00:18:16 that the sky wasn't a bowl. It was like the sun was one of these stars I was looking at. I was just hanging out there in three space and kind of an an anxiety, an angst, existential angst. And I remember this warmth when I felt like, yes, but in science, you can talk to people who are worrying about the same thing. And it's almost like a social, supportive experience. We can sort of all come together as humanity and come to the answers together. And it was kind of a warmth at that point that this wasn't just me alone sitting there hanging on a star, but it was a group. We can work together on this.
Starting point is 00:18:50 Yeah, I definitely agree that that's a really important part of how the whole scientific process works. Like the fact that there is a community that you're supporting one another. And honestly, the problems that we're trying to solve right now, and we'll talk about some of these later, are so complex that, you know, this notion that a lone genius can go do something that is, you know, like really revolutionary has always been a fiction. You know, like we're always building on what others have built before us. And in many cases,
Starting point is 00:19:27 the problems themselves that we're trying to tackle are of such vast complexity that you have to have lots and lots and lots of people working on them simultaneously in order to make real progress. So you went to Stanford. How did you decide to go to Stanford? And what was your major as an undergraduate? So as an undergraduate, everybody in my family, we all went to state schools. I think I spent an afternoon on a ping pong table filling out a form. I wasn't thinking much about college.
Starting point is 00:19:59 I just said, you know, that's what we do. I went to the State University of New York at Binghamton, which was the top school in New York when I was applying to schools. And when I got to the university, I just absolutely loved every class I was taking. And I said, I'm curious about physics and biology. Those two things were like where most of my curiosities were clustered. And so I started taking a bunch of physics classes and a bunch of biology classes, biochemistry, and so on. And at some point, I didn't want to stop looking at both. And I went to a mentor advisor whose class I loved. He taught a class in biophysics believe it or not I said
Starting point is 00:20:48 yeah this is great and I asked him I said there's no major in biophysics what do you think it would take to like do a special major in this area where I can really work with you and put together an undergraduate sequence that would really capture what you would do if you were going to study this area, even as an undergrad. And this was Professor Starzak. And we sat together and came up with a program and took it to what was called the Innovative Program Board. And a committee looked at this proposed major and they said, good to go. Now, there's a lot more classes in this and directed readings with some incredible professors. But I felt like I had the best of both worlds, chemistry, physics,
Starting point is 00:21:33 math, bio together. And then in the middle of all this, I ran into two professors as I was getting to junior and senior year. Both were just remarkable. One professor is Howard Patti, who was a professor from Stanford, actually. He did his PhD at Stanford, and his interest was emergent phenomenon. And particularly, he looked at biology from the point of view of a physicist and symbol systems. And he wrote some beautiful pieces, essays. They're still celebrated. They just not too long ago had a celebration of his career. I was immersed in Howard Patee's readings and thinkings, which were very deep and interesting
Starting point is 00:22:16 and cutting to the core of, I would call, the theoretical foundations of biology from the point of view of a physicist. And at the same time, I started talking to Robert Isaacson, who was taking a biophysics perspective on brains, looking at limbic systems in rats. I started talking to him, and he persuaded me to work in his lab, and I started looking at living neural networks. And I started getting very interested in brains. I hadn't really been thinking a lot about brains and minds since trying to build an electronic brain in like, you know, first and second grade with peanut cans and springs and wires and clay and light bulbs. So right towards the senior year, I was trying to pull together my biophysics background to looking at how brains work.
Starting point is 00:23:08 I ended up reading a couple of books. One was Herb Simon's book called Sciences of the Artificial. And another book was Michael Arbib's book, Brains, Machines, and Mathematics. And both were very motivating to me in terms of the questions that were being asked. And so I ended up applying to graduate programs which combined neurobiology with an MD. I thought, why not get into the human, have this human dimension to understand the clinical world.
Starting point is 00:23:40 Someday we'll understand brains. So I ended up getting a bunch of acceptances and had to choose among places that had up getting a bunch of acceptances and had to choose among places that had more of a mix of things and flexibility around your degree. And that would be very classical MD-PhD work and very focused. And ended up on a set of intuitions thinking through Stanford might have more of the mix that I was looking for, but I wasn't sure. Because by the time I ended up going to grad school, I was really zooming towards computation at a time where you wouldn't be thinking as an undergrad about getting into the 80s about artificial intelligence. You'd be thinking about neurobiology, neuroscience, biophysics. And so when I hit Stanford, there I was interested in getting going on neurobiology, being thrown, believe it or not, into a medical school class with a cadaver
Starting point is 00:24:32 where I ran into some close colleagues who actually had similar interests to mine. And at Stanford, you could wander off to main campus, which was just a bike ride away. So I spent a lot of time in my first year taking classes in computer science, artificial intelligence, philosophy of mind, and cognitive psychology, along with the regular medical school classes. And towards the second year, I said, you know something, I need to, I don't think neurobio is going to have the right mix for me in my pursuit of my core curiosities about what the freak was going on with minds, with human brains and the brains
Starting point is 00:25:12 of vertebrates and other animals on the planet. And that the fastest path to insights would be through computer science. And I remember one of the moments I was thinking about what I'd been doing in my laboratory work that I became very good at, doing unit recordings, looking at small circuits, listening to the ticks on the speaker in a darkened room, looking at the oscilloscope on interesting questions about how particular subsystems worked,
Starting point is 00:25:40 the thermoregulation subsystem in a rat, for example. And thinking that what I was doing all those years was sticking a thin wire into a chip and trying to infer an operating system in the application code, you know, and even the hardware by listening into the Morse code of a single gate. And I felt like that would be a waste
Starting point is 00:26:03 of good time on the planet. And I remember thinking it was a major shift to say, I'm going to give up the pursuit of a neurobiology, neuroscience PhD, and I'm going to move over now to go all in on what I came to know as artificial intelligence research, history, depth, you know, with the methods, I wanted to really master it. And it's sort of an interesting time in the 80s for AI. So one of the things we've chatted about on the podcast before is the fact that AI has had this distinct cycle of booms and busts over the years that at the Dartmouth workshop in 1955,
Starting point is 00:26:48 the program that McCarthy and these luminaries put together was way more ambitious than they, in reality, were going to be able to accomplish. And that we have had several of these cycles where the enthusiasm and the expectation for what we were going to be able to accomplish is sort of far exceeded our ability, which leads to these AI winters where you've got a bust and, you know, like people sort of go sour on the whole discipline. And as I remember it, the 80s, I forget what time in the 80s, but by the time I got to grad school,
Starting point is 00:27:32 we were well into an AI winter where it was no longer this fashionable thing in graduate programs. When you got into AI, was it right before the AI winter that we had? Or was Stanford some way a unique island where the enthusiasm for the field was undiminished over time?
Starting point is 00:27:59 I've often said that that proposal is written so well and it's so aspirational that if you submitted it today to DARPA or National Science Foundation, you'd probably get a high grant scoring to be funded. Just go for it. I mean, as written. So back to your question, when I first jumped in, it was 84. It was kind of a warm time and getting hotter. It was the time where the rule-based expert system, these production systems that do, for example, backward chaining through these modular human expert rules
Starting point is 00:28:39 were becoming quite popular. And I remember one of my first meetings was IJCAI 1985 at UCLA. And it was just an amazing time of excitement and inspiration with thousands of attendees. It felt like NeurIPS feels today. But 84, 85, 86, there was a kind of a collapse of interest and a bunch of startups going out of business that had been funded during the earlier time when it was discovered it was just kind of hard to build these systems and maintain them. And maybe these logical systems weren't as powerful and as promising as people had thought they'd be and weren't as easy to use or build.
Starting point is 00:29:21 I've looked over the history quite carefully. And what's called AI winter for us, when I say us, it was I and students at Stanford that were studying similar topic areas. And we had very close friends that we met at conferences and workshops at MIT and CMU and a few other places that created this invisible college of grad students that were looking for a different way to do things. And in many ways, we were up against the glowing cinders, and you might call them ashes, of what had been really exciting just two or three years before, typically pioneered by the people who are our mentors and advisors, which created some tension. And what we were looking for was going back to the basics
Starting point is 00:30:09 and building on the shoulders of the great statisticians and probabilists and folks who had done inference and optimization over decades. We discovered that the AI of the time in the early 80s to mid-80s was defining itself as, no, no, no, no, that's operations research. Or no, no, no, there are too many numbers there. Numbers aren't symbols. We want to make it like symbols. There was lots of tension there.
Starting point is 00:30:40 And at times, I know for me in particular I specifically I sought out new advisors at the time and moved over to working with George Danzig who was an operations research leader who if people know his work he's fabulous, but his intellectual contributions include the simplex method for optimization. Yep. And Ronald Howard, who had defined the phrase decision analysis and was really interested in thinking through how do you build systems that can help clarify thinking and bring together multiple factors under uncertainty. And what I found in George Danzig and Ron Howard, decision theorist and an optimization probabilist and folks like Brad Efron and stats, where they were looking across campus at the AI people
Starting point is 00:31:36 and thinking like, what the hell are those people thinking? And so what I started doing, I felt like, and it wasn't just me, there was a few of us in, I had a close colleague, David Heckerman, Michael Wellman at MIT,
Starting point is 00:31:55 Orin Etzioni at CMU, and others, started to think through, like, what were the big questions in AI? Even going back to the 1950s documents and before, and how could we start to build on what we knew was the science of optimization, decision-making,
Starting point is 00:32:16 action under uncertainty, high-stakes consideration of preferences, trade-offs, and started pushing in a direction that at first was considered quite foreign and outside of AI, not AI. A very distinguished professor told me that, you know, after listening to me talking about bounded rationality with using probability as the basic fabric and decision theory, he said, you know what? You have something we call physics envy, which is, I guess, referring to Freudian notions of another kind of envy. And, you know, you really need to look at symbols and high-level manipulation of predicates, go back to theorem proving. You're really wasting your time with these numerical methods. They call them numerical.
Starting point is 00:33:08 Even as we were coming up with abstractions like Bayesian networks and influence diagrams, which are higher-level constructs, representations. And I remember at the time we were joking about getting bumper stickers as grad students, rebellious grad students, driving around campus that we're going to say, numbers are symbols too. It was that bad in those days. Well, it's really interesting though because what you all were collectively doing, sort of steering things away from the symbol manipulation like systems of logic
Starting point is 00:33:39 sorts of research and getting things into this more, you know, sort of statistical framework has basically set the course for artificial intelligence over the past three decades. I mean, like most of what we talk about now when we're talking about artificial intelligence is statistical machine learning of some flavor or another. And that's really sort of a stunning thing to have that foundational piece persist for as long as it has. I don't know whether you all were cognizant of what you were
Starting point is 00:34:19 doing, but it's a really big deal that the field pivoted in that way. Yeah, rather than being cognizant, we felt like we were outsiders with some really important ideas to share. There was a panel at AAAI in 1984, I recall, where several people were almost booed off stage as we tried to bring up this idea of uncertainty in AI, principles of uncertainty. And that next year, in 1985, we decided to take that panel and make it into a workshop
Starting point is 00:34:56 we called Uncertainty in AI, UAI, which was a separate Roman community. And I remember the moment this outlandish thing happened in 2007, I was invited to be the president of AAAI. We joked, we being the former Invisible College, that the revolution was complete. And it really felt that way. Like, all of a sudden, we said, you know, look, I remember like just in like 2010 or so, like, you know, AAA, I always said, it's like UAI.
Starting point is 00:35:33 It's like UAI. It's like a big UAI now. It's everything. So let's, you know, since I want to make sure we get to some of the like really interesting stuff that we've been doing, we've been doing recently, let's fast forward all the way to some of the AI work that you've been doing over the past handful of years, which I think is of, again, really foundational importance. Like maybe even more important than this shift that you all agitated for and sort of realized when you were grad students. And that is sort of thinking about AI in the human context.
Starting point is 00:36:13 So as these technologies have become unbelievably more powerful, like especially over the past like 10 or 15 years, and their applicability to problem solving in the real world has never been higher. We are now being faced with a whole bunch of questions about what's the ethics of applying this particular algorithm or technique in this scenario? How do we make sure that these systems are doing things in unbiased ways. What is fairness in these systems? What are the things that we shouldn't use AI for? Where are the places where AI should always have decision-making systems where there should always be a human in the loop?
Starting point is 00:36:59 And so you've done, I think, some of the really most important work in the field at Microsoft and in these organizations that you have helped to start and are involved in the leadership of, like the Partnership for AI, on thinking about what the ethical responsibility frameworks are for doing AI in a modern world. How did you decide that that was something that was going to be such an important focus for you? I've always been interested in high-stakes decision-making, decisions that really make a difference in the world. This is why I wanted probability and utility theory to have a formal foundation for these actions and recommendations, applications in healthcare.
Starting point is 00:37:45 Back in the 80s and early 90s, our goal was just to get something to work. But even during those times, we saw interesting challenges with moving these systems into the actual world of usage. Like doctors who wanted to understand, like understand why the system made a recommendation, who would say things like, no, that can't be right. Can it explain itself to me? And coming up with methods to do explanation, even in the late 80s, seeing how important that would be, this human connection. I remember working on a project where I was NASA Mission Control in Houston in my last
Starting point is 00:38:24 year of my dissertation work, looking at some high-stakes decisions, time-critical decisions with propulsion control people. And I realized that it wasn't just making recommendations, it was figuring out what to display to people to help them make a decision. So there's this open-world issue that became very important to me as part of understanding the bigger role of AI in larger human settings. To me, it was more or less obvious in high stakes areas,
Starting point is 00:38:52 you had to consider these things. And then when I was become AAA president, it was a time where there was lots of initial discussions about the singularity coming, and there was both utopian and dystopian views being debated. And so I decided to make the theme of my presidency, AI in the Open World. And there's an open world of how do you put a system that's limited into more complex worlds and give it the ability to understand its own abilities and to be really much more omnipresent about the reality of helping humans out and or controlling
Starting point is 00:39:31 a system that has a function in the world not just this narrow wedge of of expertise on a certain particular classification topic or prediction and a second theme was thinking more deeply about the influences of our projects in the world. And this was 2008. And when it came to that part, I remember in my presidential address, which each president gives an address, I talked about the technical aspects and then the social and societal aspects. And I called together a group of about 25 people to create a study that I called the long-term futures of AI and its influences. And we had three groups meeting, and we ended up doing something very interestingly and analogous to the Asilomar meeting that the biologists had held in 1975. I think it was looking at recombinant DNA. But we ended up doing a workshop, a three-day workshop at Asilomar.
Starting point is 00:40:30 We all came back from our breakout groups to do reports. And it was the first time I heard this phrase from the short-term acute challenges group. We had a long-term group, a short-term problems group, and then an ethics and legal team as part of this effort. But it really drew me in and got me excited because I heard this phrase, criminal AI. I said, wow, what is criminal AI? And this group reported their findings about the malevolent use of AI by state and non-state actors and where it could go. We had another team. I explicitly asked a team to look at, could they take something as you might call fanciful,
Starting point is 00:41:11 but interesting as Isaac Asimov's laws of robotics, which folks have read about in his robot series, and actually codify them in a system with modern AI techniques so that the system could be proven to be reliable and responsible to human beings and to society. And we had a great breakout session on that topic. And then we looked at ethics and legal issues. So that whole experience and working with the community on these topics, which resonated deeply with me me for my interest in seeing these technologies do well in real hard decision problems and recommendations
Starting point is 00:41:48 further pulled me into thinking more deeply about the role that we would have as researchers, as scientists, as professionals, and as companies in thinking through not just the technology itself, which was growing in its power and its usage in commerce, as well as in areas like defense and healthcare, but to really consider deeply what we might do as a growing field,
Starting point is 00:42:15 including technical issues, social issues, looking at the human dimension. You know, if people are using these systems, how do you design them in a way not just where they might explain their reasoning if people want to know what's going on, but how do you understand how to apply the technology in a way that will complement expertise that's already available from human beings? How do you think through longer-term futures where the technology begins to shape the nature of work and the nature of the tasks people do in particularly
Starting point is 00:42:53 named jobs, like I'm a doctor, I'm a lawyer, I repair automobiles, to understand what it all would mean so that we would know how the role of humans would be co-evolving with the role of machines. And then it's just been so rewarding to see the rise of a whole field that's studying problems that are also rising in issues, for example, around the bias of systems that are trained on data that comes from cultures and societies that have all sorts of nuanced histories that lead to sampling issues and data that represents the societies from which it came and systems that you can build from that data that will amplify existing inequities in society.
Starting point is 00:43:40 So to see, you know, there's actually a rising field now of people that are pointing out these examples and coming up with ways to better visualize and understand and address them. it seems very likely to me that we are going to want to apply our most powerful technology platforms, including modern machine learning, to over the course of the next several decades. And I think perhaps the most interesting area or sort of intersection is what's happening right now in biology and how that intersects with the work that's going on in AI and high-performance computing. And, you know, it was sort of an interesting intersection and has been growing increasingly interesting over the past handful of years. And now it's just sort of acutely interesting and urgent and necessary because of all of the work that we need to do to adapt ourselves to handling the COVID-19 pandemic.
Starting point is 00:44:54 So you might not have realized how prescient you were when you were choosing to get a PhD in AI and a medical degree. But you're in this interesting position now where you have this background and point of view and visibility into this intersection. And I'd love to hear your thoughts about where you think things might be going over the next handful of years. Yeah, it's pretty impressive to see the connections
Starting point is 00:45:23 between computer science and ideas of abstraction, modularity, the ability to simulate in computer science, and where it's touching on biology. that I studied with Howard Patee on looking at biology as a physical system that had certain interesting properties that most of the world might look at as magic or in a different category, but in reality is a very interesting set of mechanisms that even relies on, for example, hierarchical abstraction the way our programs do. Yeah, so one of the things that you and I saw recently, like we went to chat with Drew Indy at Stanford, and one of the things that he said in that conversation that was so interesting to me is that now that we understand, and we're still early days,
Starting point is 00:46:24 but we understand a little bit of how to program or reprogram biology to do different things than what the biological systems do on their own. that you have yeast, which are basically little breweries. They are biological organisms that transform things like carbohydrates into carbon dioxide and alcohol, for instance. But his assertion was you've got these yeasts that we could reprogram to brew a whole wide range of compounds that we could use, for instance, as we pursue sustainability. So, like, things that might be alternatives to things that we would synthesize with petrochemicals, for instance. And, like, that's a really exciting idea. You know, we are in the early days of a major revolution in our ability to manipulate the physical world. Biology has figured this out in beautiful ways, has built
Starting point is 00:47:36 beautiful mechanisms that synthesize, that do incredible acts of chemistry, physics that are robust, self-replicating, the spinning out of shapes and structures through embryogenesis, just these magical, because you don't understand them, capabilities are coming into focus now through the lens in part of computation, physics, biochemistry. But perhaps the biggest insights in terms of the lens and the capabilities and the opportunities, I think are at the intersection of metaphors and concrete mechanisms from computer science
Starting point is 00:48:15 staring directly at biology and looking at the information theoretic aspects of what goes on in biology and then thinking about how these things can be, these systems can be harnessed in new ways, as well as borrowing ideas from biology and thinking through how we build systems, how we design materials and so on. But your comment focused more on like, how can we better modulate, moderate, design biological systems to do new acts of creation with applications in biology, applications in healthcare, applications in material science, applications in neuroscience. I think that we know so little even today, and there's so much possible that I think our
Starting point is 00:49:02 grandkids will live in a very different world where biology will have played a major role in almost everything that's being touched, including the materials that we build systems from. And so I'm very excited about the possibilities. It's bringing me back to my roots of biophysics now combined with computer science and artificial intelligence. So I'm happy to be in this new role. And as we've been talking, to start thinking deeply, working with partners like Drew Endy, David Baker, York Sea League, so many people that really are lit up now with thinking in this way of looking at what we call
Starting point is 00:49:43 a rising field of synthetic biology. How do you program biology? How do you guide it in new ways? How do you understand and control cancer? It's a runaway program. And we can just go topic by topic and think through what the engineering paradigm shifts we might need to design new kinds of robust and predictable functionalities in biological systems. You know, even something like the magic of something like the ribosome. We've all learned about the ribosome in basic biology classes. Oh, it's this interesting coalescence of RNA and protein in a structure that takes symbols and builds effectors and structures. It's one of these key, we'll call it a key pivot point of what makes biology biology you know storing up coincidences and insights in
Starting point is 00:50:47 long pieces of tape called dna the ability to take those codes as they've been learned and to transform those codes into structure and function and then to have experience re-encoded through evolutionary processes back into that tape. But this idea of a ribosome, you know, what do we understand and what can we better understand about this hinge point between information and the physics of life, I think has a lot to say about many things that we do. And, you know, the way that you just described just described that, I hope that that is one of the exciting and inspirational things that kids today will sort of see. In fact, if someone had as eloquently described genetics and the mechanism of the ribosome
Starting point is 00:51:43 when I was a high school student, as you just did, I may have chosen to pursue a different field. But I think we are at this incredibly inspiring moment where, you know, not only will our, you know, kids and grandkids live in a, you know, much different world because of what we are able to do with our new understanding of how to leverage biology to help make people healthier and help maybe make our physical world more sustainable. But I think they actually are going to be the ones who take inspiration now in what's possible, and they're going to go build this world. And that is super exciting. Yeah, absolutely.
Starting point is 00:52:26 Cool. So we are just about out of time, but I've got one last question that I wanted to ask you. So I'm always interested about what scientists and technologists who are themselves inherently curious about the world, what do you do for hobbies? What's a thing that people might not know about you? That's interesting. Well, I try to get exercise,
Starting point is 00:53:00 and people might not know that I'm a hacker, which is I just stopped playing, but I've been playing ice hockey in the Greater Seattle Hockey League for a number of years on a team called the Hackers. It seems that most people, most of our opponents think that we're actually of hacker on the ice. So I find it's one place where I turn off everything except really focusing on teamwork and where the puck is and being out of breath and how fit I am. And I've loved those kinds of sessions, being out on the ice. I just decided to step off the team when I was getting busier and busier.
Starting point is 00:53:50 And I was actually one afternoon very grumpy at an all-hands meeting at Microsoft. And I just come back from a game from Everett at two in the morning. And I decided, you know what, I just can't do this anymore. So instead, I took up inline skating now, believe it or not, you know, but I wanted to be serious about it. So I, during NeurIPS, for example, this year in Vancouver, I went to this custom blade shop and I had like these marathon blades made and I committed to being in the Berlin inline skating marathon on September 26th. And just last night, I was worried about this. They popped up a message saying they've canceled it and they'll be in touch with us. But I was
Starting point is 00:54:31 training for the Berlin inline skating marathon down the streets up and down here with these new, with these Vancouver blades that are just like magnets on the asphalt. But I do like to get out and get my mind focused on just clearing it with running or skating or paddleboarding, other kinds of things I enjoy reading. Just coming off a really interesting book. I tend to read Science Magazine every week, and there's a great book review they do of books coming out in the sciences. And I just finished this book called Becoming Wild by Carl Safina, which is looking at animal culture, looking at, for example, sperm whales. And it's just really amazing to read about the interaction of the importance of culture, stuff that's passed down
Starting point is 00:55:28 among animals versus being in their genetic code for thousands of years, tens of thousands of years, and different cultures, even with the same species living side by side, different dialects that are spoken by whales, for example. I've always been interested in the, this gets into the AI and the open world question, but even brains in the open world, how did human beings, how did our minds, our nervous systems co-evolve with our culture, co-evolve with tools like language? And so I found lots of interest there
Starting point is 00:56:02 in that recent book that I read with core questions that I have about the role of our external world and our tools like languages with the shapes and operation of homo sapien nervous systems. That's awesome. So this has been a great conversation, Eric. Thank you so much for taking time out to chat with us today. And I really, as usual, enjoyed hearing more about what you're thinking. Yeah, great catching up, Kevin. I'm looking forward to continuing our discussions. Awesome.
Starting point is 00:56:37 All right, so that was Kevin's chat with Eric Horvitz, who is Microsoft's first chief scientific officer. I loved hearing his story about his journey getting into tech and medicine, which is such an interesting combination, as well as your broader discussion about ethics in AI and where that future holds. Yeah, I think one of the remarkable things about Eric, and this has been true for so many of the guests on the podcast and so many of the people that I have the good fortune to know working in technology, is that one of the things that motivates them and that has motivated them since they were really young children is this
Starting point is 00:57:19 voracious curiosity and desire to understand what's going on in the world. And it was really, really great to hear Eric talk about that part of his life. And, you know, I think part of the reason why he chose to do things the way that he did them in terms of how he approached his education and how he has spent most of his research career is because he just refused to decide on one thing or the other. He's like, why can't it be all of the above? And that's really what we need more of when we're thinking about AI,
Starting point is 00:57:57 especially as this technology is having a greater and greater impact on society and our future. No, I couldn't agree more because it brings a really good kind of way of looking at the world that you might not get otherwise. Yeah, I mean, one of the really fortuitous things about having Eric here at Microsoft and having him play such an important role at Microsoft over the years is that when a moment like now arises where we really do have to think more comprehensively than ever before about what this intersection is between biology and artificial intelligence,
Starting point is 00:58:33 it's pretty convenient to have one of your foremost AI experts actually be a medical doctor as well. It's sort of only at Microsoft, I guess. It really is. And I mean, you didn't really get into this too much in your conversation, but I just want to ask you, especially since you've written so much about AI and since you've been having these conversations with people like Eric who are experts in biology, what role do you think AI might be able to play kind of going forward as we're looking at how to combat this
Starting point is 00:59:06 and maybe even other potential viruses or health concerns? Yeah, I think we're seeing it have a really tremendous impact already. So, you know, as we have dug in with a bunch of the researchers and a bunch of the medical professionals and biotechnologists over the past handful of months, it's already the case that they're using the tools of machine learning and artificial intelligence in relatively sophisticated ways.
Starting point is 00:59:33 So it may be on one end of the spectrum using natural language technology to better extract critical information out of our unstructured health records that are, you know, for many, many, many years now have been handwritten notes or, you know, notes that are taken and input into a medical record system. But it's still, you know, sort of this unstructured data that we really do need to be able to establish more structure around so that we can do the types of deep analytics that we need to do to understand things. like SARS, coronavirus too, and really try to widely disseminate what effective therapies are
Starting point is 01:00:28 that people are applying over time. And so, funny enough, natural language processing and natural language understanding, which are these classic techniques from artificial intelligence, have huge relevance there. You also have seen the work that people have been doing, and we've talked about this some on the podcast, in using deep neural networks to do medical diagnostics. And so I'm wearing a ring from a company called Aura right now that measures your body temperature and your pulse and a whole bunch of things about your movement. And I think this company originally intended to have these rings help you manage your all-in health, like whether you're sleeping well enough or whether you're getting enough exercise and activity. But it may be the case that the data that's gathered by devices like this are going to be really useful when you are able to train sophisticated deep neural networks with them in detecting diseases like COVID-19,
Starting point is 01:01:33 hopefully before you're gravely ill and have time to go get yourself treated so that you can jump back to robust good health as quickly as possible. And then on the very far end of the spectrum, which has been some of the most surprising bits for me to see over the past handful of years, is how the tools of AI, in particular deep reinforcement learning, are almost becoming like a new calculus for the basic sciences.
Starting point is 01:02:08 So you had calculus come about as this analytic framework for better describing and understanding the phenomena in the real world in the 18th century. And you got most of modern science from having a tool like that and I'm seeing now with AI deep neural networks and machine learning deep reinforcement learning
Starting point is 01:02:34 these new self-supervised learning techniques that we've developed over the past handful of years are being applied in science in sort of the same way that you might imagine calculus was many years ago to more accurately and faithfully model the phenomena in the physical world so that you can better understand them.
Starting point is 01:02:51 And that might be helping to accelerate a molecular simulation that's trying to understand how the spike glycoprotein in the coronavirus is interacting with your epithelial tissue and invading cells and infecting you with this horrible disease. And we are already seeing how machine learning and AI, these new techniques are being used to accelerate those simulations and to get to more accurate results.
Starting point is 01:03:24 So I think there's going to just be almost like a landslide of activity and building momentum over the next handful of years as these two worlds, artificial intelligence and biology, start to intersect in a more profound way. And I think we're going to spend a bunch of time this season on Behind the Tech talking to some of these innovators who are in the biosciences using these tools and these innovative ways to help make us all healthier and bring better healthcare outcomes to as many people as humanly possible and to use biology in ways that we really weren't even conceiving of a few decades ago.
Starting point is 01:04:09 That's great. I'm glad. And what's great about this, I think it gives us hope and we need hope right now. I know that I'm certainly feeling hopeful. Well, that's a wrap for us today. As always, please reach out anytime at BehindTheTech at Microsoft.com. We'd really like to hear from you.
Starting point is 01:04:26 How are you faring during these times? We'd love for you to share some of your stories about how you're innovating, how you're hacking, and finding ways to stay connected with technology. We'd love to hear from you. Thanks for listening. See you next time.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.