Microsoft Research Podcast - 020 - Getting Good VIBEs from Your Computer with Dr. Mary Czerwinski
Episode Date: April 18, 2018Emotions are fundamental to human interaction, but in a world where humans are increasingly interacting with AI systems, Dr. Mary Czerwinski, Principal Researcher and Research Manager of the Visualiza...tion and Interaction for Business and Entertainment group at Microsoft Research, believes emotions may be fundamental to our interactions with machines as well. And through her team’s work in affective computing, the quest to bring Artificial Emotional Intelligence – or AEI – to our computers may be closer than we think. Today, Dr. Czerwinski tells us how a cognitive psychologist found her way into the research division of the world’s largest software company, suggests that rather than trying to be productive 24/7, we should aim for Emotional Homeostasis instead, and tells us how, if we do it right, our machines could become a sort of “emotional at-work DJ,” sensing and responding to our emotional states, and helping us to become happier and more productive at the same time.
Transcript
Discussion (0)
We were calling it like a DJ. It's your emotional at-work DJ. So, you know, when you need everything
really ramped up and going strong, like a DJ knows how to do, but then when it's time
to calm the crowd down, maybe our software can learn, okay, Mary needs to take a break.
Mary needs to slow down. Mary needs to go get a glass of water.
You're listening to the Microsoft Research Podcast, a show that brings you closer to
the cutting edge of technology research and the scientists behind it. I'm your host, Gretchen Huizenga.
Emotions are fundamental to human interaction, but in a world where humans are increasingly
interacting with AI systems, Dr. Mary Cherwinsky, principal researcher and research manager of the Visualization and Interaction
for Business and Entertainment group at Microsoft Research,
believes emotions may be fundamental to our interactions
with machines as well.
And through her team's work in affective computing,
the quest to bring artificial emotional intelligence, or AEI,
to our computers may be closer than we think. Today, Dr. Cherwinsky tells
us how a cognitive psychologist found her way into the research division of the world's largest
software company, suggests that rather than trying to be productive 24-7, we should aim for emotional
homeostasis instead, and tells us how, if we do it right, our machines could become a sort of emotional at-work DJ,
sensing and responding to our emotional states and helping us to become happier and more productive at the same time.
That and much more on this episode of the Microsoft Research Podcast. Mary Trewinsky, welcome to the podcast today. It's great to have you with us.
Thank you. It's an honor to be here.
So you work under the larger umbrella of human-computer interaction at Microsoft Research, but more specifically, you're the principal researcher and the research manager of the VIBE group.
Right.
Tell us what VIBE stands for and what gets people in your group up in the morning.
VIBE stands for Visualization and Interaction for Business and Entertainment.
Most of the people in the group are very deeply into various aspects of information visualization,
so helping people work with big data better or building tools for programmers
so programmers can deal with all the vast amounts of data we have coming in these days, or in the area of effective
computing, which is where I'm squarely putting my research these days.
Microsoft Research has a number of social science researchers now.
Yes.
But you actually were the first social science researcher.
How did you end up here?
Did they come looking for you? Who decided they needed a Ph.D. in cognitive psychology to help with computer science research? having social scientists on their research teams. But it was actually Eric Horvitz and George
Robertson that actually sought me out as a psychologist to partner with them in their
research around attention and rendering on the screen and redesigning windows in 3D.
And so the two of them really kind of helped usher me over to research. I was actually in
product at the time. So I was in Microsoft, just wasn't in the research group. So it was a great time to come out into the field as a cognitive psychologist because
back then we had one computer screen and it was all about users being able to look at
the screen and make sense of it.
Well, that's what cognitive psychologists study, perception and cognition around information,
learning, memory, making decisions, attending to various
aspects of the displays. So it was perfect timing for me to put what I did my PhD in
to work for Microsoft. You know, I would have not even gone there in terms of that thinking.
Just given where we are today with the attention economy, they're calling it. And there's so many more demands on our attention. So tell me
how it's evolved since when you started to now. Yeah, well, I mean, back then, the experiments we
would do, for instance, with the astronauts at Johnson Space Center, I would be looking at their
layouts of their displays, and they'd be running nine experiments at a time, and they had to
monitor them all the time, in addition to the various space station systems like thermonuclear control, which is kind of important. And so we
looked at character-based user interfaces and how people can track that kind of information.
That all changed when graphical user interfaces came along. And then it was more about how you
grouped things and how you arranged them on the screen. And it's still about that today. That hasn't gone away. But now, you know, everything's more social and now we have multiple devices and
multiple displays we have to look at. So it's just fanned our attention out even more broadly
and harder to control, obviously. Yeah. So it makes your job a little harder as well.
It does. Yes. But that's exciting. Yeah, it's really fun.
Your research focuses on a couple
of big ideas, emotion tracking, information worker task management, and healthcare and wellness for
individuals and groups. And we'll get to each of those in turn. Okay. But for starters, let's talk
a little bit about this big idea of affective computing and the quest for artificial emotional
intelligence or AEI. What is it, and why is it important?
Well, I don't think we'll actually have quite natural user interfaces until we can actually
talk to our systems the way you and I naturally talk to each other. And so a large part of that
is these really natural, almost automatic social signals that we cue off of each other.
So the way you're nodding your head
gently in agreement with me is the way that I know that you're mimicking me basically and my thoughts.
So we believe that systems will need to have that kind of emotional intelligence, EQ, in order for
them to be natural enough for us to really engage with them and want to keep engaging with them.
So there's not that freaky, uncanny valley where it just doesn't feel natural, doesn't feel right. So we've been pursuing this emotion tracking in order to track your emotion
as a user so that the system can respond more like a human would emotionally, appropriately.
So do you think we'll ever get real artificial emotional intelligence with a machine?
Well, that's TBD. I would rather be on the edge
exploring those things as they happen as a scientist and trying to make the most of it.
Part of your research involves designing what you call delightful systems,
but it isn't always easy to tell whether they're really delightful or not. How are you tackling
this, Mary? I like to say, how can we say that we're designing systems
that delight our users if we don't actually know? So that is why we are using cameras and microphones
to look at our users' faces, to see what their facial gestures are. Are they smiling? Are they
squinting their eyes? Are they frowning? But also listen to their voices, hear how they're saying
things about working with our systems. And if we can build these kinds of
emotion sensing platforms into all of our products, then we'll know from the get-go if we're actually
designing systems and tools that users are delighted by, or are they frustrated, and then we
should fix those areas that seem to bring on the frustration. So I think if you build these tools
into the platform of all of our systems that we build in the future, we'll know better if we're making systems and software that people love.
You were part of a recorded presentation called, And How Does That Make You Feel?
Which I loved the name of.
It's like the psychiatrist.
Does that make you feel?
Exactly.
And at several points, you got frustrated with the fact that your video wouldn't play.
And about the third time,
you actually flashed a big smile and go, that's just great. And I remember thinking,
how would I as a machine know that you weren't really happy and delighted with me?
That's a great question. That's a great question. Because we never think,
now I'm channeling Daniel McDuff on my team. We never believe that a one to 10 second slice of a user's face is going to give us anything worthwhile. What we believe in is that you have
to watch and listen to this user for a long time, longitudinally, so that we really understand the
range of a person's emotions. People are all over the place with how labile their emotional state is.
I'm pretty up and down.
I have a broad range.
But some people aren't.
Some people are much flatter or just more controlled with their emotions.
And so you really do have to study an individual for a long time before you know.
But then also, the machine needs context about what you're doing.
So if the machine had been watching and seeing that I couldn't play my video, the machine would learn that that smile is probably a smile of complete frustration.
Or a grimace.
Or a grimace, a grimacing smile. It could have been. But yeah, so the machine needs the context
and the longitudinal data about you, yourself as a person, before it can have really good
assuredness in that it's making the right emotional classification.
So that's an interesting thread to drill in on because you want to build a system that will
work for many people, not just one person. And if my emotional ups and downs are different from
hers, then how are you going to bring the research together and make a machine that
can tell the difference if you need to track me for a long time and him for a long, you know, times a million. Yeah, well, because you can do
this now with simple webcam and mic on any laptop or PC, it's pretty easy and the user doesn't have
to do anything to do this kind of tracking longitudinally. And we're doing it right now
here in Building 99. Many of us are getting our emotions classified while we're using everyday
software systems and doing our regular work.
So now we have thousands and thousands of hours of these emotions tracked, and we can bring them back to bear, not only with generalized models, but also personalized models that could be more relevant just to me.
So we can do both scales quite well.
Okay, so you could even employ machine learning AI to actually customize in a particular product for a particular person.
Yes, yes, that's the angle.
That would be cool.
Mm-hmm.
You framed some of your research in the category of productivity tools.
Tell us about the technology you're working on to help mitigate negative interruptions, multitasking, or I think you really would call it task switching because there is no such thing.
At least that's what I tell my daughter.
You're not multitasking.
And some of the general lack of focus that we experience at work and if we're honest in life. at managing their workflow. So they stay pretty concentrated and focused for longer periods of
time than other people. And they take breaks when they know that they need to, when they need to
refresh, because no one can stay focused 24-7. That's a myth. So we like to refer to something
called emotional homeostasis, which is you stay focused, you work hard, maybe you are even a
little stressed, and then you do something to balance that out, right? So maybe you use social media if you're one of those people. But other people have
a harder time staying focused for longer periods of time. And actually, every time they hear a ping
or every time they get an email notification, they go to it. You know, the little numbers on
the icons are the worst thing for these kinds of people because they see they have something new
and they have to go check it. So for those kinds of people, we can actually use machine learning and this personalized kind of software we've been talking about to see when you need to be in the flow, to see when you actually do your best work and to kind of help you stay there.
So perhaps that's the time when you're really starting to get into your work.
Perhaps that's the time we turn off notifications or hold less high priority notifications away from you. Turn off that inbox, you know, dinger and let you focus.
We might even turn off social media if we go so far. The user gives us permission to do that.
We'll do whatever the user wants us to do to keep them in the zone. And then, you know, when you
tend to kind of come out of that flow, maybe it's right before lunch. Let it all flow back in.
It'll be just fine. And the machine can see that you're moving out of the zone and it's going to
turn back on automatically? This is current research we're doing right now. This is our
hypothesis that we'll be able to see when you prefer to have your hardcore work moments and
when you are okay with letting notifications through and possible work breaks through.
Or we can see if you're one of those people that stays really focused, maybe we can see you've been focused a little too long for yourself. Maybe you are getting a little stressed
out and a little walk would be quite welcome at that point, but it's all work that's in progress.
So we still don't know how these, we kind of like, we were calling it like a DJ. It's your emotional
at work DJ. So, you know, when you need everything really ramped up and going strong, like a DJ knows how to do. But then when it's time to calm the crowd down, maybe our software can learn, okay, Mary needs to take a break. Mary needs to slow down. Mary needs to go get a glass of water, bio break, something, you know. So we'll see how much users will be welcoming such software and will design it until they do welcome it,
will iterate it and do it better. If we're using these tools to get in the zone and stay there,
should we call it artificial productivity? No, it's real productivity.
I'm kidding. It's artificially assisted. Artificially assisted productivity.
Yeah, yeah. Well, so let me ask you this as a psychologist,
what happens to good old-fashioned
self-discipline? Have we just lost that in this digital world? Well, I will tell you, in a study
we ran two summers ago, a couple of our participants, we turned off all their social media,
all their notifications, they weren't allowed to use their phone for eight hours a day. And a couple
of our participants did make note of the fact that they had lost sight
of how often they were going to social media and checking their phones. And they were shocked by
how much more time they had. So in some sense, the sad truth is, yes, we have trained ourself
to do this task switching on an almost constant basis. And on a great note, software can help.
So, you know, maybe we can train it back.
Let's get a little meta on that topic and this whole concept of digital assistance and affective agents.
And kind of on the same note, have we really reached the point where we need technology to save us from our technology?
I just did a press interview yesterday where I did actually make the claim that we need technology to help us from our technology.
So I can stand by that. I've said that. But as I said, all of our tasks switching
bad behavior was trained. It was trained by the tools and the features that we ask software
companies to build for us or tech companies in general. So we can use the technology to retrain
ourselves to focus. I believe we can. I know we can. The research shows we can. And we'll make better decisions. We'll do better at our jobs and we'll be more productive.
So yes, if we have to start using technology as training wheels to get rid of all the bad
technology that's, you know, bifurcating and dividing our attention, I don't think that's
a bad thing. I think it's a good thing. We've come around full circle. From a psychological point of view, aren't we at the risk of outsourcing some basic human skill sets that are emotional, that are necessary for emotional growth and health?
Right. So I wouldn't say that we've gone that far, but it is nice that we can use the machine intelligence to kind of protect us again and get us back into focus. I think that's going to be great. In terms of the emotional side of things, I really do think that
the advent of robots taking care of children and the elderly, it's almost upon us. That is an
outsourcing, in my personal opinion, possibly a necessary one, but it's still an outsourcing.
And so that's why I'm really, really vehemently opposed to not
studying what the effects of children using robots and personal assistants is on their own
communication, growth and behavior. But also I'm really adamant that machines need to have EQ,
not because I don't want you to know that you're working with a machine or a robot,
but because I want that conversation to have an emotional balance and to be emotionally mature so that kids don't grow up with this imbalance, with lack of EQ, lack of real intelligence.
Right.
So I think it goes both ways.
And that's why I want to study both of those aspects of it, because it could be worrisome, right?
Absolutely.
So on that thread, affective
computing, as I'm looking at it, seems to have two sides to it. One is designing machines that
can interpret human emotional states and adapt to them. And the other is designing machines that can
help humans interpret their own emotional states. First, am I right? And second, if so, why is this two-sided approach
necessary and happening? Most people aren't very aware of their emotional state. And so what we
find when we first start doing our experiments, it actually takes people a day or two to actually
start understanding how they actually feel. I'm usually in the positive quadrant of how we do
these self-ratings. And it took me even a
little while to realize I'm not always, you know, happy and high energy, right? Sometimes I am kind
of low energy and sometimes I am sad. So after you track yourself for a while and you are honest
with yourself, then you start to realize how the states move around. And so in any experiment,
we always know that at the very beginning, users might not be super good at it, but they'll get better at it over time. And I think that it is very useful because we forget our emotional states very, very quickly. Quick little survey we did showed people forget how they were feeling about a day. And if it's a really big event that happened in your life, you're probably going to remember how you felt. It's those little patterns of micro badness and micro goodness we don't actually pay that much attention to. And it could really help us make better decisions. Like for
instance, people were telling us after like a week of tracking themselves and whatnot, that they
wouldn't remember these things a year from now. So having a system that tracked it would be quite
useful. Now it has to track it accurately. So there should be a way that you can correct the
system when it's wrong. But I think it's kind of nice to go back and look at how I was feeling. And actually,
I'm always surprised I am happier most of the time and frustrated very little of the time.
But there are moments you would forget otherwise. So it's useful to you. It might be useful
if you were taking care of a loved one to be able to see those tracks. Right now,
most people go to a therapist's office or a doctor's office. They fill out a paper form
for how they've been feeling the last four weeks.
Well, people don't remember how they were feeling the last four weeks.
So these tools can be very useful if they could be shared with loved ones and caregivers.
Yeah.
So let me ask a question about that.
If I'm going to be monitored in all these situations, what's watching me?
What's recording?
What I want to know is I'm in front of my
computer right now and it's a laptop and I might be working at a desktop at my desk, but out in
the wild, I don't have these devices. Oh, right. Now you would have to have some device that has
a camera and or a mic, but you can have either one that's constantly listening to you in those
particular contexts that you might want it to.
Right now, it's very easy when people are just at their desk or in a meeting.
So actually, we're tackling meetings next, which will be fascinating.
Oh, wow.
Yeah.
But yeah, you would have to have some kind of recording device on you if you wanted to get it 24-7 all day long.
Well, let's talk about that for a second, and then we'll swing back to some other topics.
Increasingly, the mechanisms that are collecting data include video, audio, maybe wearables,
ingestibles, implantables, lunchables, I don't know.
How do we reconcile the desire to have this knowledge, this quantified self, this self-awareness
with our desire for privacy
and our fear of big brother.
Great.
No, that's why we do everything we can to keep your signal private to you.
There's no identifiable information that goes into the cloud at all.
We as researchers can't go in and look at your data.
If you're running our system, we can't.
And so is it encrypted?
I mean, how would I be confident?
You are completely de-identifiable.
We can't find you unless you give us a code back, you know, your ID and tell us you want
your data, which of course we can give you if you want it.
That's the only way we know who you are.
Well, I just interviewed Kristen Lauder, who works in cryptography. And it's like all of the different teams that come together to make these tools both fantastic and trustworthy at the same time is super important.
And we will surely use her stuff.
We work with Ronnie on her team, so he talks to me about what they're doing all the time.
Oh, good.
It'll be a good thing to have.
Yeah.
In fact, you used a funny phrase to frame the human-computer
interaction group's relationship to other groups at Microsoft. Tell us what it is and why you say
it. Yeah, human-computer interaction, people often play the glue between various technology teams
because we're usually the user-facing part of the experiment. And we can see when, you know,
there's a technology that's very right to present to a user, but it needs something. Like we were just talking about, it needs encrypted privacy. So we'll bring the two
teams that need to talk to each other together because we go out and put our user interfaces
on all of their technologies. So we kind of know what they're doing. So that's why I say that.
Sure. So tell me what's going on in the research stage and even in the product stage on things that can help me track my emotions or my state of being at work. What we do know is that users really could use
help from our technology to turn off notifications, to possibly turn off non-work-related
websites like social media, and help the user focus. And maybe the user just wants to say
something like, I just want to turn all that off for a half hour. Then I'll be fine. And the system should help them do that. Eventually,
the goal would be to do that automatically for the user. So the user doesn't have to remember
to do something manually on their own. That's an interesting thing. I don't know if it's better if
I develop the skill to remember to turn it off or to have somebody do it for me.
We have to do studies on all of this.
Right.
Right.
One of the biggest problems I have, and I know I'm not alone,
is wrangling and making sense of my data. How are you helping me?
Ah, well, I know you did a podcast of Stephen Drucker. So he's one of the people in our group
that does wonderful work in information visualization. But we have many researchers
that do work in information visualization at Microsoft Research. And they're all fabulous.
They all take different aspects of how you look at your data. Some of them like to work with networked data. Some of them like to work with time-based data. So there are many, many tools in the InfoViz community. InfoViz stands for information visualization.
Thank you.
You're welcome. That can be thrown at these kinds of big data problems.
Yeah. Danielle Fisher has done in our group, which I thought was really nice, is he starts to show you
your data as it's coming in. But if there's so much data that it can't all be rendered at once,
it just gives you a feel for what the data is going to look like. And you decide when you think
you have enough to make a decision. So they're all working with these kinds of interactive
visualization tools to help you see patterns in your data that you might not have been able to see
in an Excel spreadsheet using standard bar graphs or line graphs.
So you said some of this is still in the research stage, but some of it's already incorporated into BI and Excel and things like that.
Yes. Power BI has been a great vehicle for taking visualization work and exposing it to end users.
And I think it's been very helpful.
It's very hard to come
up with novel visualizations. So it takes a lot of creativity, a lot of work to cook up something
that people can use pretty quickly because they're usually pretty sophisticated versus the standard
bar chart or pie charts. So it does take users a little while to learn how to use them, but then
when they do, they're very powerful and useful. Yeah, yeah. Let's look to the future for a second. Your group is publishing prolifically. You're going to CHI in a bit with a lot of papers.
What excites you most about what the field is doing right now and what might be exciting to
the next generation of human-computer interaction scholars? Right. Well, I think the whole movement towards the gig economy
is really new and exciting, and it's studied pretty hard at CHI. It's represented pretty
well at CHI. So I tend to go into two tracks when I'm at CHI because there are like 14 parallel
tracks. So you have to pick. And I tend to go into mental health and e-health. And then I try to go
to the gig economy tracks because that world is moving so fast. So the advent of micro tasks are emerging. You know, even just think of
writing a Word document now. We have researchers in the building who are working on just taking
little parts of writing that document and making it a task for somebody to do out there, right?
Jamie Teevan and I did a whole podcast on this.
Okay, good. And Shamsi Iqbal is doing it as well as some others. But that economy, whether you're talking about taxi drivers to airplane designers, that economy is going gig. And so it's just fascinating to listen to hear the research they're doing and the trends that are happening and how those gig workers actually work with each other to make sure each other succeed and they make a good wage. So to me, I always follow those two tracks.
And I think the young guns will need to do a lot of work on the gig economy.
So that's exciting.
But also health and mental health.
I mean, really moving fast and using machine learning to look at, you know,
precision psychology, for instance, is a huge topic.
So I'm hoping the young people look at those two areas.
But there are many more very hot areas. Those are just my two favorite. We talked about what gets you up in the morning.
Is there anything particularly that keeps you up at night, Mary? Yes, I do get worried about
our systems that will eventually have good EQ, but they don't yet. and how working with the systems today when they're not truly intelligent
they don't filter by age gender location etc in terms of what they say pretty much these days
I worry that it changes people's communication patterns because they're going to pattern off
the system so I really get worried in particular about young people the elderly I imagine their
morals are already built in and they've learned their communication style pretty well by now. But I really worry about the young
generation and the generation that maybe, as we said, maybe they're going to be working more with
a robot than they are with, you know, a teaching assistant in the future. So I want to make sure
that that modeling is what we as a society see as appropriate, moral, and positive. So I really want to focus on
generational issues like that. And also what is the appropriate EQ for a machine, right? Do you
want to know that it's not human? I think most of us want to know that it's not human. And so
how do you do that and yet still make the conversation feel natural and make the conversation
feel appropriate? That question is so profound
because what much of the work happening here is about is making computers more human-like.
And yet, when you ask that question, do you want to know? I'm nodding my head really hard. Yes,
I want to know it's not. I don't want to be fooled by the Turing test. Right, right. So we have to, again, as a society, come up with a plan, basically.
Are you doing that?
Thinking about it hard.
If you weren't a researcher, what would you be?
I really wanted to go back and get my PhD in biopsychology at one point.
I, you know, I can't not be a researcher. I'm sorry.
I was just going to say, you answered that by saying I'd be a different kind of researcher.
I mean, if I could be a professional tennis player, I'd do it, but I don't have the talent,
so it's not going to happen. Last question. What advice would you give to your 25-year-old self?
This is for the researchers out there who would be looking at,
what do I do? Where do I put my time, talent, treasure? Yeah, what I tend to tell young people
that I mentor that are still in graduate school, for instance, because that's about 25, I tell them
not to think that life is just a straight line, right? It's never a straight line. You don't get
your PhD and go from point A to point
B to point C to point D. Sometimes if you stay in academia, that can happen. But even then, I don't
think your career is going to go in a straight line. Too much happens. And in my particular case,
technology changes way too fast. So you always have to be open to something. I thought for sure
I was going to go into academia.
I couldn't because of a two-body problem. I ended up in industry. Oh my God, was that the best thing for me ever? That was just perfect. And then I jumped around, right? I went to Johnson Space
Center. I went to Bell Communications Research. I kept my foot in academia because I thought,
you know, someday I might want to teach again. Got this wonderful job at Compaq,
which led me to Microsoft, which led me to Microsoft Research.
You know, you just got to stay open and you'll know when it's time to leave. Things won't feel right. And then that great opportunity might be out there and just be open to it.
Mary Cherwinski, thanks for joining us today and sharing so much about what you're doing. And
we can't wait to see what's going to happen.
Thank you. It was my pleasure.
To learn more about Dr. Mary Cherwinsky and how to have a better relationship with your computer,
visit Microsoft.com slash research.