Microsoft Research Podcast - 030 - Examining the Social Impacts of Artificial Intelligence with Dr. Fernando Diaz
Episode Date: June 27, 2018Developing complex artificial intelligence systems in a lab is a challenging task, but what happens when they go into production and interact with real humans? That’s what researchers like Dr. Ferna...ndo Diaz, a Principal Research Manager at Microsoft Research Montreal, want to know. He and his colleagues are trying to understand – and address – the social implications of these systems as they enter the open world. Today, Dr. Diaz shares his insights on the kinds of questions we need to be asking about artificial intelligence and its impact on society. He also talks about how algorithms can affect your taste in music, and why now, more than ever, computer science education needs to teach ethics along with algorithms.
Transcript
Discussion (0)
If I'm running an experiment, what is an ethical experiment?
What's an unethical experiment to run on users?
To what extent should users be aware of the fact that they're in experiments?
How am I going to recognize and address biases in the data that my machine's learning from?
And that's just scratching the surface.
There's going to be plenty of other questions that are going to be raised in the next few years
about how we design these systems in a way that is respectful of our users.
You're listening to the Microsoft Research Podcast, a show that brings you closer to
the cutting edge of technology research and the scientists behind it. I'm your host, Gretchen
Huizenga. Developing complex artificial intelligence systems in a lab is a challenging task.
But what happens when they go into production and interact with real humans?
That's what researchers like Dr. Fernando Diaz, a principal research manager at Microsoft
Research Montreal, want to know.
He and his colleagues are trying to understand and address the social implications of these
systems as they enter the open world.
Today, Dr. Diaz shares his insights on the kinds of questions we need to be asking about artificial
intelligence and its impact on society. He also talks about how algorithms can affect your taste
in music and why now more than ever, computer science education needs to teach ethics along with algorithms.
That and much more on this episode of the Microsoft Research Podcast.
Fernando Diaz, welcome to the podcast.
Thank you.
You're a principal research manager at Microsoft Research Montreal, and you work in artificial intelligence and search and information retrieval, but you're also really involved in fairness, accountability, transparency, and ethics or FATE research. So in broad strokes, and we'll get specific in a bit, what gets you up in the morning?
What are the big questions you want answers to and the big problems you'd like to solve?
Well, a lot of these systems that we're building are extremely successful.
So information retrieval or web search, computer vision, these have all been developed over the course of many, many decades.
And they're starting to get productionized and users are starting to see them on a daily basis.
What we haven't thought so much about in designing these systems is, as computer scientists, what is the social context in which these things are being used?
And so my concern here is better understanding, you know, what are the social implications of these systems we're building? How will the social context in which these systems are used affect not just our own metrics like precision or recall, but also society at large?
And I think this is something that's really coming to a fore for computer scientists because
a lot of these techniques that have been developed in isolation are just starting to be
commercialized right now. So you're a computer scientist by training, and you've done research
in information retrieval, statistical methods, machine learning. But you've lately become very
interested in the interface of these AI systems and society, particularly what happens when AI
hits production, or as some people say, the open world. Why this interest for you now?
What questions are you asking? What piqued your interest in going this direction? That's a great question. So I obviously went to graduate school and got a PhD
and was studying these systems at a pretty abstract level and really conducting experiments
with data that was static and collected offline. And soon after graduate school, I came into
industrial research lab and was working with production teams that were implementing the techniques that I was studying in graduate school.
And you begin to realize that when you take these algorithms and you scale them out and you put them in front of real users,
a lot of the basic assumptions that you were making in the laboratory were not well supported in reality. And so that was sort of a check for me in terms of my research
agenda, really coming back to the first principles and trying to understand better, okay, what is the
problem here? What exactly do I need to be measuring? And how do I optimize for this right
metric, whatever that might be? So you were at Microsoft Research before, and then you took a
bit of a break, and then you came back. You started in
New York, and now you're in Montreal. What brought you back? So I had actually, after graduate school,
started my industrial research career in Montreal. And for various reasons, I had to relocate out of
Montreal. But even while I lived here, I recognized that this city itself, and Canada in general,
has a pretty rich tradition
of very strong computer science research, machine learning research. And in the back of my head,
like I'd always wanted to come back to participate in that. And so when the opportunity arose to
rejoin Microsoft Research in a new Montreal lab, it was the perfect fit, especially since
the lab itself focuses on artificial intelligence.
The city itself is really going through a blossoming of AI research and being part of that and contributing my voice to that broader conversation is something that just made sense to me.
Let's talk about Montreal for a minute.
It's become a global hotspot in artificial intelligence research.
And the goal of the MSR Montreal lab is very specific.
They want to teach machines to read, think, and communicate like humans.
Give us your particular take on where we are on this quest and how your research interests align with what the Montreal lab is doing.
Well, I think part of the reason there's a research lab in that area is the fact that there are still a lot of unanswered questions with respect to how to engineer these systems.
And I think it will require more than just folks from natural language processing or more than just folks from dialogue or reinforcement learning.
It really requires the combination of those.
And I think that's one of the things that makes this lab especially unique. Now, my role is to come into this lab and hopefully add to that conversation by providing a perspective of, well, how will these systems actually behave when they're in a human context?
Because, like I said before, it's very easy to design these systems in isolation.
And then when you deploy them, realize that there were some very strong assumptions in
the experiments that you were conducting. Now, the role of the group that I'm building is to try to
sort of anticipate some of these questions and better engineer our systems to be robust to, say,
differences in the populations that I might be interacting with or in the corpora that I might be
gaining knowledge from. What is the group you're building there?
So the group that I'm building is sort of a sibling organization
of the Fairness, Accountability, Transparency, and Ethics group
that we started in New York City a few years ago.
The focus will be on studying the social implications
of artificial intelligence in society.
And so that group will be composed of folks with a technical
computer science background, but also a technical background from related disciplines such as
sociology. And so the idea here being that in order for computer scientists to better
understand and address the social implications, they really need experts in, you know, sociology. They need experts in anthropology,
et cetera, to give us the insight into those things that we have not been measuring so well so far.
Yeah, let's talk about that. The fairness, accountability, transparency, and ethics
application to a variety of artificial intelligence and machine learning technology
research is super important right now.
And as you bring up, this is because not all coroner cases can be explored in the lab,
and there are some unintended consequences, along with the intended ones, that can surprise people.
And so this community doing research in this area is quite diverse in terms of academic training.
What do each of these experts bring to the mix when they're looking at fairness,
accountability, transparency, and ethics? So somebody from a social science background
will have a better understanding of technology use in general, how it's used or misused,
and how people just react to certain tools that we provide them. Somebody from a legal background
might be able to better comment on the policy implications
of certain technologies that are being developed or really give us a deeper understanding of what
we mean when we talk about something like fairness. And then folks from the computer
science community really understand the systems that are being developed and may be able to
conceptualize some of the constraints such as fairness and incorporate
them into the system. But it really requires these multiple perspectives to come up with
better approaches to designing these systems. Let's go back to some stuff you've done in the
past and are still working on now, information access systems and search engine information
retrieval. And in a paper you wrote, you suggested there's
a gap between studying these systems and implementing them, but you also make a sort
of provocative statement that there are open problems that are better addressed by academia
than industry. What kinds of problems are those, and why do you make that statement?
One of the things that happens in information access research is you have academics who have contributed to the community a lot, but these days a lot of the, say, web search research is happening at the big web search companies where they have the data, they have the users, etc.
And a lot of times the academics don't have access to the experimentation platforms or the data.
And so there's a disparity in terms of the amount of rigor you can do in your research.
So what I was claiming in that article was that, you know, well, academics, even though they don't
have that amount of data, they do have a broad set of collaborators that you may not find at
the bigger search engine companies. So at a university, you have access to sociologists
in other departments, you have access to economics, professors, all of these are potential collaborators, which will help you understand
the problem from multiple different perspectives, instead of perhaps one very specific perspective,
which you might have in a web search company. I think data set releases are one strategy.
One of the other approaches that I think are one of the other types of scientific platforms that one would
not have in academia is the experimentation. So I can actually run A-B tests, you can run
controlled experiments with large populations of users, which doesn't really exist in a dataset
release. No, that's true. And so one of the things that I think is worth exploring is how do we
actually provide access to academics for doing that sort of
controlled experimentation. Interesting. That's happened in bits and pieces here and there,
but I think this is really something that we as an industrial researcher can think about providing. Okay, let's go back to data. Let's talk about data for a minute. In machine learning circles,
there seems to be some consensus that it's no longer good enough just to have big data,
and I use that with air quotes around it, but it also has to be good data or unbiased data,
which is part of that mix. So we know that there's probable bias in a lot of these big data sets,
and we need to fix that. People are now talking about trying to remove bias through things like
search engine audits and fairness-aware algorithms and that kind of thing. How do you do that?
One of the reasons we're concerned about bias in data is that the trained model will be biased when it's deployed. And so step one is to
be able to detect whether or not the actions of artificial intelligence are biased themselves.
And if they are, how do I go back and retrain that algorithm or add constraints to the algorithm so it doesn't learn the biases
from the data. And so my work to date has focused primarily on the measurement side of things.
On the measurement side of things, it has more to do with understanding the users that are coming
into the system, what they're asking for, and whether or not the system, by virtue of the fact
of who the user is or what population they're coming from, is behaving in a way that you would consider biased.
And that requires a lot of the expertise from the information retrieval community who have been thinking a lot about measurement and evaluation for almost since the beginning of the research agenda of the community in the 50s.
And so this is what makes it a good natural fit between auditing and measurement and information retrieval.
So as we've discussed, bias has become a bit of a buzzword in data science and AI research.
But you've suggested there are other social issues besides bias that also need to be addressed. What
are some of these issues and how can your research help?
Yeah, I do think that bias is a very important problem.
But I think one of the reasons why I talk about the social implications of AI is because I think bias is just one of the social implications that we can detect.
There are certainly others.
So a pretty clear one is transparency. So how do I make the algorithm's decisions about what it's doing transparent to the user so that the user feels a little bit more in control of the situation when they're actually trying to cooperate with an algorithm?
A second one would be sort of the cultural implications of algorithms.
So this happens more in context of, say, a movie or music recommendation.
So I'm building this big system to recommend music to individuals.
What are the longer-term cultural implications of deploying these recommendation algorithms if I
know that recommending certain musicians will push somebody's musical taste in certain directions
in the long term? What does this mean for the creation or curation of culture?
The other side of that problem is that music recommendation
algorithms can really have profound effects on the creators and musicians themselves.
And so I might, as a computer scientist, say, well, this is the best algorithm for music
recommendation. I'm just going to deploy it. But as computer scientists, we haven't really
thought about, well, what are the effects on the actual creators? I think that, for me,
is one that is especially salient. So on that thread then, how might you craft a research
project that would ask these questions and how would you do your research there? Right. So let's
take this example of music recommendation. So you can imagine sitting down with musicians and
better understanding understanding what is
important to them. What do they feel like they're getting out of a system? How do they feel like
they're in control or out of control in a recommendation system? And sitting down with
folks who come more from the sociology or anthropology backgrounds, media studies,
to help me as a computer scientist understand what that big population of musicians look like. And then I can, as a computer scientist, sit down and try to
better understand, well, how do I design an algorithm which will both satisfy my listeners
as well as satisfy my musicians? Now, I think even posing it that way is extremely reductive.
And so that's why I wish there was somebody in the room from one of these other disciplines to point out and say, well, Fernando, you know,
you haven't thought about this. So given the nature of the research you do and all the things
you see as a result,
is there anything we should be concerned about, anything that keeps you up at night?
One of the things that does concern me is the fact that a lot of these techniques that we're developing as a research community are being put out and then deployed into production within a
matter of days or weeks by people who perhaps were not the original experimenters. Now,
there's nothing wrong with open science, but I do think that we need to be a bit more cognizant
and aware of the effects that our algorithms will have before we rush to deploy them.
And what I'm concerned about is that we're sort of quickly pushing out newer and newer algorithms
without having a deeper understanding of those implications. Microsoft Research has a reputation for working closely with academic institutions,
and I know that education is something you're passionate about. So talk about what's going on
in FATE, or Fairness, Accountability, Transparency, and Ethics, education, and tell us your vision for
the future of education in this area. So I think when I went to graduate school or even undergraduate for computer science,
one of the things that was not really taught so well was ethics or the social implications of the
technologies that we're developing. And I think part of the reason for that is that, you know,
you're studying operating systems or you're studying information retrieval at an abstract
level. You're not really thinking about the context in which they're going to be deployed. And to be honest, the class is hard
enough having to just think about the core algorithms. And so I think at least when I was
trained and even now, like understanding the social implications and social responsibility
that you have as an engineer or scientist did not really make it into the conversation. And so I think that's
being recognized now. I think that computer science departments are starting to develop
curricula around ethics and computer science. And so I think students are starting to be trained.
My concern, though, is that, you know, we already have a lot of people out there developing systems
that have not gone through that training. On top of that, there's not a lot, I mean, you don't need a computer science PhD in order to,
you know, start up a company. So that part of the, you know, who's not covered by education
is another thing that keeps me up at night. But in terms of the education side of things,
I think as somebody who's been in an industrial research lab, I can provide that perspective to students when I'm in the classroom.
So to better understand not just the practical implications of deploying a machine learning system at scale, but also the social and ethical implications of deploying a machine learning system at scale.
And this is exactly the sorts of questions I alluded to before.
If I'm running an experiment, what is an ethical experiment?
What's an unethical experiment to run on users? To what extent should users be aware of the fact that they're in experiments? How am I going to recognize and address biases in the data
that my machine's learning from? And that's just scratching the surface. There's going to be plenty
of other questions that are going to be raised in the next few years about how we design these systems in a way that is respectful of our
users. You know, you see the history of other areas, for example, medicine and the Hippocratic
Oath, first do no harm, and also the systems that are set up in place to help ensure that people
aren't harmed. And we think naturally about that with physical trials. And
it's just really encouraging to know that we're starting to think about that in artificial
intelligence deployment at large scale, even though we don't have like an FDA for AI, there's movement.
Yeah. And I think it's important to understand that that movement has happened in
other disciplines, as you said. And so it's not like we're going through this for the first time.
So to the extent that we can reach out to folks in the medical ethics community or folks in the
social sciences, they can help us develop what it means to be a responsible computer scientist.
I mean, if we were starting from ground zero, that's especially daunting. But I think we do have collaborators across the disciplines that
can help us with this problem. So as we close, Fernando, what's on the horizon for the next
generation of researchers in your area? I know that's a big question, but at least what thoughts
or advice might you leave with our listeners, many of whom are aspiring researchers, who might
have an interest in the social impact of AI? So I think for folks interested in that area
specifically, I think a lot of the research that's happening right now is really day one of a much,
much longer research agenda. I think it's an extremely exciting area to be involved in because
while we're starting to ask the fundamental questions now,
I also think that there's a lot of additional fundamental questions that have yet to be asked by exactly those young researchers right now.
I think in terms of the social impact of that research, it's potentially very large because there are a lot of these systems that are already deployed. And so your ability to be involved in, say, a product like a search engine that has social
impacts and to correct those things is extremely powerful.
Yeah.
So it's a great time to be a young researcher.
Fernando Diaz, it's been great talking to you and super encouraging.
So thanks for taking time to talk to us today.
Thank you very much.
To learn more about Dr. Fernando Diaz and his research on fairness, accountability,
transparency, and ethics in computer science, visit microsoft.com slash research.