Microsoft Research Podcast - 009r - Life at the Intersection of AI and Society with Dr. Ece Kamar

Episode Date: July 3, 2019

This episode first aired in January, 2018. As the reality of artificial intelligence continues to capture our imagination, and critical AI systems enter our world at a rapid pace, Dr. Ece Kamar, a sen...ior researcher in the Adaptive Systems and Interaction Group at Microsoft Research, is working to help us understand AI’s far-reaching implications, both as we use it, and as we build it.Today, Dr. Kamar talks about the complementarity between humans and machines, debunks some common misperceptions about AI, reveals how we can overcome bias and blind spots by putting humans in the AI loop, and argues convincingly that, despite everything machines can do (and they can do a lot), humans are still “the real deal.” https://www.microsoft.com/research

Transcript
Discussion (0)
Starting point is 00:00:00 When I first talked with Dr. H.A. Kumar in January of 2018, we covered a wide range of topics surrounding artificial intelligence, including the technical promises, common misconceptions, and societal challenges that AI researchers are facing today. Unsurprisingly, they're all still relevant a year and a half later. Whether you heard H.A.'s description of the AI landscape at the beginning of last year, or you're just arriving on the scene today, I know you'll enjoy episode nine of the Microsoft Research Podcast, Life at the Intersection of AI and Society.
Starting point is 00:00:33 Machines are good at recognizing patterns. Machines are good at doing things in scale. Humans are good at common sense. We have creativity. So I think if we can understand what this complementarity means and then build AI that can use the power of AI to complement what humans are good at and support them in things that they want to spend time on, I think that is the beautiful future I foresee from the collaboration of humans and machines. You're listening to the Microsoft Research Podcast, a show that brings you closer to the cutting edge of technology research and the scientists behind it. I'm your host, Gretchen Huizinga. As the reality of artificial intelligence continues to capture our imagination and critical AI systems enter our world at a rapid pace, Dr. H. A. Kumar, a senior researcher in the
Starting point is 00:01:32 Adaptive Systems and Interaction Group at Microsoft Research, is working to help us understand AI's far-reaching implications, both as we use it and as we build it. Today, Dr. Kumar talks about the complementarity between humans and machines, debunks some common misperceptions about AI, reveals how we can overcome bias and blind spots by putting humans in the AI loop, and argues convincingly that despite everything machines can do, and they can do a lot, humans are still the real deal. That and much more on this episode of the Microsoft Research Podcast. A.J. Kumar, welcome. Thank you. Hey. You're in the Adaptive Systems and Interaction Group at Microsoft Research. What do you do there? I'm really passionate about AI systems
Starting point is 00:02:27 providing value for people for the daily tasks. And I'm very interested in the complementarity between machine intelligence and human intelligence and what kind of value can be generated from using both of them to make daily life better. If we try to build systems that can interact with people, that can work with people, and that can be beneficial for people.
Starting point is 00:02:48 Our group has a big human component, so we care about modeling the human side. And we also work on machine learning, decision-making algorithms that can make decisions appropriately for the domain they were designed for. So we work on dialogue systems, work on location-based services, but also the core principles of how we do machine learning and reasoning.
Starting point is 00:03:14 So what do you do particularly? My main area is the intersection between humans and AI. I'm very interested in, as I said, AI being useful for people because my PHA advisor, Barbara Gross, says this very eloquently. She says, we already know how to replicate human intelligence. We have babies. So let's look for what can augment human intelligence, what can make human intelligence better. And that's what my research focuses on. I really look for the complementarity in intelligences and building these experience that can in the future hopefully create superhuman experiences. So a lot of the work I do focuses on two big parts. One is how we can build AI systems that can provide value for
Starting point is 00:03:58 humans in their daily tasks and making them better. But also thinking about how humans may complement AI systems. One of the biggest misconceptions maybe about AI is that it's more advanced than it actually is. I think there's been a lot of conversation around that. Where are we with AI, really? Yeah, I think when people think about AI, they actually think AI is this thing that emerged in the last 10 years or so. I think there are a lot of misconceptions about this. AI actually started in 1950s, and there was the Dartmouth Conference on AI where the fathers of AI came together, then they actually met to discuss what AI is,
Starting point is 00:04:37 define it, and start working on it. And they were wrong, too, in their perception of what AI is. They actually, they wrote this Dartmouth document where they stated that if they worked hard on this AI problem for a summer, they really believed we could achieve AI very, very quickly. So this is a problem. Artificial intelligence is a problem where seeing a few very good examples lights up your imagination. It increases your expectations and think that, okay, if the machine can do this, I can imagine machine doing this harder thing
Starting point is 00:05:10 very quickly as well. Unfortunately, that's not how AI works today. AI actually works on verticals. We know how to apply some specific techniques to some tasks, but unfortunately, just being able to do, let's say, perception doesn't make it easier for us to do other cognitive tasks in machines. So that's a big misconception we have to correct.
Starting point is 00:05:32 We are actually very far from generalized artificial intelligence. Our techniques still are very specialized, and applying AI to every single domain requires a lot of analysis, deliberation, and careful engineering. Talk about the blind spot problem in machine learning. And where does it come from? How do we prevent or mitigate it? This is a problem that we got very passionate about in the last few years. And the reason we are very passionate about this problem is because we are actually at an important
Starting point is 00:06:12 point in the history of AI where a lot of critical AI systems are entering the real world and starting to interact with people. So we are at this infliction point where whatever AI does and the way we build AI have consequences for the society we live in. And when we look at our AI practices, it is actually very data dependent these days. So our models are data hungry. We give data to these models and the quality of the performance we see from the models at the end of the day depends on how well that data is collected. However, data collection is not a real science. We have our insights, we have our assumptions, and we do data collection that way. And that data is not always the perfect representation of the world. This creates blind spots. When our data is not the right representation of the world and is
Starting point is 00:07:06 not representing everything we care about, then our models cannot learn about some of the important things. For example, a face recognition system that cannot recognize dark-skinned people is an example of the blind spot problem. Our data may not have enough representation for these subpopulations, and then our systems do not have enough representation for these subpopulations. And then our systems do not work equally well for all people. Yeah, I think you mentioned an example of a machine seeing just pictures of white cats and black dogs. And then when it sees a black cat, it thinks it's a dog because that's all it's been fed and told. That's all it's been fed. How do we deal with the problem of flawed data or incomplete data and make it better? In the age of AI, we have to reinvent what our development pipelines look like.
Starting point is 00:07:56 In software engineering, software engineering is a decades-old field. And with the hard way people have learned what it means to troubleshoot and debug this code that they are developing. We are at this point that we have to have the same responsibility for the AI systems we are developing as well. That's a problem that my research focuses on. In the last few years, we've been developing debugging and troubleshooting technologies where we can actually smartly inject human insights into AI systems to understand where AI systems fail, where our blind spots are, or in integrated AI systems, how errors may propagate through a system, and we may not have any idea where the errors are coming
Starting point is 00:08:38 from and the ideal ways of fixing them. So the approaches I work on really look into combining algorithms with human intelligence and think about how humans may guide us in places where they are not happy with the performance of our system and show us clues about how they can be improved. I think one of the other biggest misconceptions about AI is that machines don't make mistakes. And they do. You bring up this concept of humans in the loop. Talk about that a little bit more. So for a lot of the tasks that we are designing AI systems, humans can actually be guides for us
Starting point is 00:09:16 because a lot of things like computer vision, speech recognition at this point, humans are the real deal. They can do these tasks very well. So that's what we do in these systems when we talk about human in the loop. We think about how to bring humans into training, execution, and troubleshooting of these systems as teachers. What does that look like? Because I'm envisioning research and all data comes from humans. I mean, in the sense that we feed it
Starting point is 00:09:45 into the machines. When you want to bring people into the process, what does that look like? So humans are already part of the process in AI. So one thing I like to say is AI is developed by people, with people, for people. And what I mean by that is AI is not this magic that it just emerges itself. We have engineers, scientists working in AI systems, putting their insights into the design and architecture of these systems. When I talk about development of AI with people, as I said, a lot of the resources for AI comes from people producing this data. And in this space, actually, the field of crowdsourcing, the field of human computation,
Starting point is 00:10:30 has actually been like this hidden gem for the developments we are seeing in AI. Because what crowdsourcing does through marketplaces like Mechanical Turk, they provide scalable on-demand access to human intelligence. You can push these tasks into the internet and people come and do these tasks for you. The same is social media. People produce all of this data that AI systems can learn about. And these web tools for accessing human intelligence is very important for us when we are building human-in-the-loop systems,
Starting point is 00:11:02 because now we think about putting the human intelligence through these micro tasks into the execution of systems for troubleshooting, for training, for execution, so that we can do these tasks better. And when I talk about building AI for people, a lot of the systems we care about are human driven. We want to be useful for human. So humans are actually very important also in evaluating how the systems we build are. And that's another area
Starting point is 00:11:29 where crowdsourcing comes into play. H.A., you serve on a number of program committees and study panels, and one of them is AI100, which is a 100-year study of artificial intelligence. And it has a report titled Artificial Intelligence and Life in 2030. What are the big takeaways of that for people who may not read that dense paper? Reports, yeah. Yeah, that's reports. We actually have a nice summary at the beginning of it, so I'll give you five minutes.
Starting point is 00:12:02 I would actually encourage everybody to read it. So let me tell you a little bit about what AI 100 is, and then I'll tell you our main insights from the report. AI 100 is a project started at Stanford, actually through very generous contributions of Mary and Eric Horvitz, to study the impact of AI on society. And what makes AI 100 really special is that it's a longitudinal study. It's going to continue for 100 years or more. Because right now, everybody is excited about AI. We can do one study and forget about it in the next five years. It's not what this study is about.
Starting point is 00:12:39 This study is going to repeat itself in every five years. So what we studied in AI 100 in 2016 with 17 other AI leaders is understanding the consequences of AI for daily life. There was one big consensus. We really didn't see much evidence for either general intelligence or the machine achieving some kind of consciousness in a way that we should be worried about right now. But we also identified areas where we have to be more careful about. For example, we realized that we are not paying enough attention to how we can build AI as team members for people,
Starting point is 00:13:19 how we can build collaboration into the AI systems we have today. We also got very worried about the real world short-term impacts of current AI practices on our society. So we identified issues around biases, fairness, explainability, reliability as real issues we should be worried about. Starting from today, we could actually see that these issues are coming into play. We were able to identify examples that we were worried about. So we actually made a recommendation to attract more research and interdisciplinary attention to some of the problems coming in this space. What other disciplines might be applicable or important in this research? It's actually a very interdisciplinary problem. It touches law because...
Starting point is 00:14:13 Huge. Hugely. Because we are talking about responsibilities. We are talking about when something goes wrong, who is responsible for that? We are thinking about AI algorithms that can bias their decisions based on race, gender, age. They can impact society. And there are a lot of areas like judicial decision-making that touches law. And also for every vertical, we are building these systems. And I think we should be working with the domain experts from these verticals. We need to talk to educators. We need to talk to doctors. We need to talk to doctors.
Starting point is 00:14:45 We need to talk to people who understand what that domain means and all the special considerations we should be careful about. Do you know of any effort to sort of link arms with other developers and designers and companies to have this sort of ethical approach? There is an exciting effort going on right now called Partnership on AI, where a lot of big companies who are practicing AI, including Microsoft, came together because they recognize our responsibility in building reliable, trustworthy, unbiased AI into the real world that can help people. And these companies are now together working on best practices and other kinds of awareness they can create to start addressing these negative consequences that AI may have in the real world. I think this is a very promising
Starting point is 00:15:38 step. We are going to see what comes out of it, but I think it is good to see that there's a consensus among big AI companies to come together and recognize that this is actually an important time in our field to start asking the hard questions. Yeah. And you would hope that people would be aware that there's always unintended consequences of any particular technology. So it's also that right now, the public perception is very important. And the current experiences we are putting into the real world are affecting public perception about AI. There is an important feedback loop
Starting point is 00:16:16 we have to think about so that the experiences humans have from our AI is actually a good one. And it's going to promote more work in the space rather than creating a cloud of fear. You're doing work here. And you have out in sort of the general public a lot of movies and TV shows and books, even sci-fi, that are painting a picture of AI and giving people actual perceptions of what it can do? And how do we help educate about where AI is and what it can do in the real world? It's an important question, how we can reach the public.
Starting point is 00:17:00 Because, you know, as a researcher, I publish in academic venues, I talk to my colleagues, so we can have more discussions in our specialized fields. However, the public perception is very important. This is why when we were writing the AI-100 report, one of our target audiences was the public. So we wrote everything in a way that any person can read and understand our thoughts. And that was very intentional because academic people can get access to a lot of thoughts on the space. And our main targets were the public, the regulators in this space, and the press. So we really wanted to make sure that our thoughts, this is actually our thoughts, right? This is not the field of AI, but these 18 people who spend a year working on this problem. At least our thoughts, this is actually our thoughts, right? This is not the field of AI, but these 18 people who spend a year working on this problem.
Starting point is 00:17:47 At least our thoughts, after careful considerations and deliberations, would be accessible to everybody who is interested about this. I think there are also things we can do as we are building our AI to make things more accessible for people and also get their feedback in. So we know that our AI is not always right. So if we give ways for people to tell us when they are not happy about something, and we can actually have feedback loops to watch for feedback, I think that's a way we can kind of incorporate more of the public thoughts about the way we are building these systems. So where would you implement that if you were going to get feedback from the public?
Starting point is 00:18:32 It could be as simple as like thumbs ups and thumbs down. Like I didn't like what this system didn't work for me. I didn't like what you were doing. Because if you get that kind of a feedback, we can have human eyes looking at those cases and see what we are getting wrong. Right now, for a lot of the systems we have, there is no feedback loop. We are not really getting much human feedback about what our systems are doing. We care about how we can get a little bit more accuracy from our existing data sets. Like how can we go from like 98% to 98.5? How we can do better stochastic gradient descent
Starting point is 00:19:06 and how we can optimize all of the parameters we are doing in the way we are doing supervised learning. However, we are spending so little time into watching what these systems do in practice in the real world. That's an area that my research focuses on very much these days because I think it's a very important problem. Watching what our systems are doing with real people in the field, what are their blind spots, what are the errors coming from, and what the performance in the world really looks like,
Starting point is 00:19:34 because the real world may really not look like your existing data sets in the lab. So let's put our focus a little bit more into the real world performance and making sure that our stuff works in the real world. And I think for that, we need a feedback loop for the real world performance. Talk a little bit about this delicate balance between self-governance and more external locus of control governance. How do people manage to do the right thing up front so that somebody doesn't have to come around and make a law that says you can't do that? I think that is something that the AI community is realizing so that we see efforts like partnership on AI. As a community, we are coming together and saying,
Starting point is 00:20:28 let's get things right before somebody forces us to do so. Because we see regulations coming from different countries all over the world. We saw some regulation coming from Germany on autonomous vehicles. We see regulations coming from European Union on data privacy and protection. And sometimes in some cases, if the regulators are not working hand in hand with AI experts, these rules may not be the best things for the field, even for the real world performance.
Starting point is 00:21:03 I mentioned issues around biases, for example. If we are not collecting demographic data about our users, it becomes impossible to detect how we are impacting different subpopulations and actually realize whether we have biases in real world performance. So I think it is very important to have dialogue between different people, communities who have a stake in this game, regulators, the press, the public, domain experts, to get them talking with each other. Because if we cannot explain where we are as a field to the regulators, they are going to be making some decisions that are going to be actually impacting things in reverse ways, like the ones we see in biases. And we are trying to communicate, I was talking at the Aspen Forum in the summer about jobs and AI and trying to express that, no, not all of our jobs are going away. Some tasks may be automated, but we are very far from automating a lot of the jobs that humans make. So just trying to express what we think the opportunities and the real challenges that are in our field in clear ways, I think is going to be very important as we go along.
Starting point is 00:22:18 I'm trying to wrap my brain around how we do that. I mean, getting an audience that's engaged in this topic, aside from being just entertained by this topic, that's going to take some thought and research in itself, I think. I think we have an opportunity with this new interest in AI. We have eyes on us. People are watching what companies are doing in this field of artificial intelligence. We see articles every week talking about how AI is going to be changing everything, all the sectors. Let's use this attention to make sure that our practices of AI is actually good. And we are collaborating with the people we need to collaborate with to make sure that we actually have good AI systems in the real world. Let's turn this attention into an opportunity. That's what I would say. I love that. So you're professionally
Starting point is 00:23:11 and probably personally interested in the societal impacts of AI. We've covered that, particularly as it affects what I want to call the three Ds, disruption, displacement, and distribution. Talk a little bit about those problems. It is one of the things we studied in the AI 100 report as well. And it's the number one question I get when I talk to people outside of AI, like, where are the jobs going? Can you tell me how many millions of jobs are going to be lost in the United States in the next 10, 20 years, and how many new jobs are going to be coming. We are actually quite far from having general artificial intelligence. Because of that, we really don't see this kind of singularity like AI happening, where AI is going
Starting point is 00:24:00 to kind of teach itself to become smarter and smarter. And one day it's going to be so smart that all the jobs are going to be going away. At the same time, we have an opportunity to think about how to prepare our society for some of the job displacement that may happen. For example, that crowdsourcing marketplaces I was telling you about, like how we are using that to train our AI. That's actually a good way to provide work to people who don't have traditional work right now. So that's an economy unto itself there, a little micro economy. It's a micro economy. And we've actually done some experiments. We have research paper on it where we actually show that we can have algorithms that can sequence tasks in crowdsourcing marketplaces so that by doing tasks, people actually learn how to do complex tasks better.
Starting point is 00:24:55 So we can train people for tasks without them even noticing that they are getting better at tasks by just doing them. At Microsoft, we have a platform like LinkedIn that can actually be a hub for providing work or for training people or for making aware of other opportunities they may not be aware of. So I have a feeling that if you can think about how to prepare generations for upcoming opportunities and pair that up with platforms we have today. Maybe that's an area we have to focus on really hard so that as some automation is coming in, we are actually preparing the people for the changes they may face. Not all jobs are going to go away because as a person who is working in AI, I just know how hard it is to automate
Starting point is 00:25:45 even the smallest of tasks. So we are going to be seeing a transition, but I don't buy the statements that all jobs are going to be going away. And not only that, we've seen this before and we see new economies grow out of change. What things do you see that might paint this picture of the human-AI collaboration that you talked about earlier? I think the key is understanding the complementary strengths of people and machines. We know that machines are good at automating repetitive tasks that we have a lot of data for. Machines are good at recognizing patterns. Machines are good at doing things in scale. But there are a lot of things machines are not good at. Humans are good at common sense. We can do counterfactual reasoning. We are not even
Starting point is 00:26:40 close to what this means for machines. We can learn from very few examples and generalize that to things that we've never seen before. We have creativity. So I think if we can understand what this complementarity means and then build AI that can use the power of AI to complement what humans are good at and support them in things that they want to spend time on. I think that is the beautiful future I foresee from the collaboration of humans and machines. We are already seeing examples of this. There is this great paper from, it's called the work between MIT and Harvard.
Starting point is 00:27:18 They can actually have a machine do nearly as well as human radiologists on diagnosing breast cancer from images. So that's a story that we hear a lot. But the story we don't get to hear a lot is because machines and humans make different kinds of mistakes. Then you actually put them together. The error rates you get is much, much lower from what the machine and humans can do alone. They can actually reduce human mistakes by 85% by just pairing them up with people. I think those kind of stories about like how humans and machines can think about problems in different way and complement them, they are going to be ways for us to just, as humans, get better at what we are doing today. What should we be concerned about or interested about in our ability
Starting point is 00:28:12 to impute human qualities to machines, form emotional connections to machines? Yeah, we've done experiments in my PhD days, and we were actually like showing the same algorithm to people. In one case, we were telling them it was a person. In the other case, we were saying it was an agent. And the humans were treating them very differently. Really? Yeah. But then Cliff Nass has some work that shows that even putting like a smiley face on an agent makes people to treat that agent differently. So perception is very, very important. You know, in some cases, it may actually be beneficial to have an emotional
Starting point is 00:28:52 connection with a computer agent. For example, ShowWise, Microsoft's agent ShowWise. A lot of people in China connect with ShowWise and maybe it helps them with their daily life and just to connect with somebody because they may not have, you know, somebody to talk to every day. So in some cases, I think it's really, again, application dependent. In some cases, it may be beneficial to form an emotional connection. How did you end up at Microsoft Research? I'm actually one of those that started as an intern and got to be a full-time. So I'm originally from Turkey, did my undergrad in Turkey, then came to Harvard for my PhD. And I think I was
Starting point is 00:29:35 in my second year and I was doing work on human-computer collaboration in particular about attention management, how we can build AI algorithms that can understand the human state and actually know when is the right time to ask a question or interact with the person. And I was reading a lot of articles coming from Microsoft Research, in particular from Eric Horowitz and his team. And I just loved the way those articles were written and the applications and the way they were approaching those problems, I want to do an internship. So I came here for the first time as an intern in 2007.
Starting point is 00:30:13 And I just loved it so much. I came back again in 2008 as an intern. And then I became a Microsoft Research Fellow. And I just love the place so much, the freedom, the ability to work on real world problems and the opportunity to do impact. Those were the things that really made this place the best of both worlds for me. The fact that you could work on real world problems in scale, but also be in touch with academia, publish, share your work, learn every day about new problems. So when I was graduating, it was a very easy decision to come here. And I've been here for the last seven plus years. It's hard to believe time goes so fast.
Starting point is 00:30:54 But I've been in the same group, working with the same manager all of this time. Ece Kumar, thanks. Thank you. It was fun. To learn more about Dr. H. A. Kumar's work and what Microsoft is doing to make artificial intelligence reliable, unbiased, and trustworthy, visit Microsoft.com slash research.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.