Microsoft Research Podcast - 093 - Data science and ML for human well-being with Jina Suh
Episode Date: October 9, 2019Using technology to help us improve our health is nothing new: a quick web search returns hundreds of apps and devices claiming to help us get fit, quit smoking, master anxiety or just “find our cen...ter.” What is new is a serious cohort of researchers exploring how artificial emotional intelligence, or AEI, could help us understand ourselves better and, when used in concert with human caregivers, enhance our well-being. One of those researchers is Jina Suh, a former Xbox developer who got hooked on research and is now an RSDE in the Human Understanding and Empathy group at MSR, as well as a PhD student in computer science at the University of Washington. On today’s podcast, Jina shares her passion for creating technologies that promote emotional resilience and mental health; gives us an inside look at an innovative research collaboration that aims to improve collaborative care for cancer patients with depression; and tells us an emotional story of how, on the brink of quitting her job, she found inspiration to get back in the game and begin a new career in research for human well-being. https://www.microsoft.com/research
Transcript
Discussion (0)
There's a huge opportunity, an unmet need of our technology being more understanding of us and us being more understanding of what the capabilities of the technology are so we can have this rich, meaningful experience collaborating with the technologies and working with the technologies. So in that sense, making our technology more emotionally aware
is a key part in making that experience meaningful.
You're listening to the Microsoft Research Podcast,
a show that brings you closer to the cutting edge of technology research
and the scientists behind it.
I'm your host, Gretchen Huizenga.
Using technology to help us improve our health is nothing new.
A quick web search returns hundreds of apps and devices claiming to help us get fit, quit smoking, master anxiety, or just find our center.
What is new is a serious cohort of researchers exploring how artificial emotional intelligence, or AEI,
could help us understand ourselves better and, when used in concert with human caregivers,
enhance our well-being. One of those researchers is Gina Su, a former Xbox developer who got
hooked on research and is now an RSDE in the Human Understanding and Empathy Group at MSR,
as well as a PhD student in computer science at the University of Washington.
On today's podcast, Gina shares her passion for creating technologies that promote emotional
resilience in mental health, gives us an inside look at an innovative research collaboration
that aims to improve collaborative care for cancer patients with depression, and tells
us an emotional story of how, on the brink of quitting her job,
she found inspiration to get back in the game and begin a new career in research for human well-being.
That and much more on this episode of the Microsoft Research Podcast.
Gina Su, welcome to the podcast.
Thank you.
You work in an interesting space, Gina.
Your title is Research Software Design Engineer.
You're currently working in a group called HU, or the Human Understanding and Empathy Group at MSR.
And you're beginning your doctoral studies in computer science at the University of Washington. In broad strokes, tell us what motivates you to do the work you do and what
gets you up in the morning. What gets me up in the morning is actually an alarm clock.
No, but really it's my daughter. She's really my inspiration. I get up to do what I set out to do
because I want to make the world a better place for her. I want to show that, you know, being a working mom is hard, but it's really rewarding.
And that learning never stops.
So I want to show her that you can redefine yourself at any age, under any circumstances.
So I do work in technology for human understanding and well-being. So I think in terms of mental health
and human well-being, I think all of us at some point go through experiences where we feel like
we're not in control of our emotions or our lives. We go through very stressful moments. We go
through very sad moments. I think there's a huge opportunity for us to develop technologies that helps us
understand who we are, what are the feelings that we're having, and really help us reflect and be
the better version of ourselves. Where is it that we would need technology to help us be better
versions of ourselves or to overcome obstacles in our emotional and mental health?
I think in a lot of ways, we have the capability to kind of look at what we've done and reflect
on our past and figure out why we're feeling certain ways. But that takes effort and motivation
and training and practice too. So a lot of people need professional help through therapies to develop the skills in order to reflect.
But I think technology, especially with multimodal sensing technologies, we have the opportunity to bring information to the person and help them look at their behaviors in ways that human memories cannot fathom.
Right. In some ways, what I'm hearing you say is that if we can set a timer or an alarm for some
physical thing we have to do, why couldn't we have technology prompt us on things that we have to do for our mental health.
Right.
And a lot of our mental health influences our productivity and our work and our thinking and our decision making.
So I think it's absolutely necessary for us to schedule for ourselves to reflect and also practice the skills that we need to do to reflect and be mindful, just like we practice for work and for exercises.
Well, we'll get to the specifics on that in a second, but let's go back upstream a little bit.
You have a dev background, but you say now you only code when you need to.
When do you need to?
Well, when I have an idea.
Okay, cool.
I was wondering about that.
You've also done some work in machine learning and HCI,
and you say you're a proud alumnus of Patrice Simard's machine teaching group.
Tell us a little bit about your previous work and how it's informed the work that you do now.
So I started as a developer in Xbox.
I worked in Xbox for seven plus years and i always admired the design team and
loved working with them and you know trying to understand the rationale behind their design
choices and i wanted to do more high-level thinking you know be closer to our customers and
you know build things that are more meaningful and have genuine reasons why we're building them. So I stumbled across something
called HCI, or human computer interaction, or HCD, human centered design. And at the same time,
I stumbled across this machine teaching group, which claimed to be this focus on like studying
the people interacting with machines. And, you know, when I started working in the group, I
also started my part time master's program at University of Washington in human-centered design and engineering.
So at first I was just helping the group build their machine teaching tool.
And then I got to help build prototypes for user studies and then help run the studies and
eventually design studies using the knowledge that I was gaining from my master's and also through work.
And I just fell in love with the science of it all and really observing people using the tools in the way that you hypothesize.
So slowly and maturely, I started doing more independent research, forming my own projects and contributing to research.
And I'm pretty much hooked now.
I am a firm believer of the machine teaching principles, you know, this idea of empowering the human teacher.
And I believe I can take those learnings and apply in other domains.
Right. And some exciting ones, too.
I want to talk a bit more in depth about this human understanding and empathy or HUE group.
Tell us in more detail what's going on in Hue and who's doing what right now.
The mission of the Hue team is to really empower people by creating and inventing new technologies
that promote emotional resilience and well-being. It's really grounded in the fact that emotions are
fundamental to human interactions and they influence everything that we do, starting from
learning, memory, decision-making, and all these other aspects of our lives. So, you know, how do
we bring emotional intelligence to technology is kind of the core of our research. Tell me a little
bit more about this idea of emotional intelligence and emotional resilience, because this isn't
something we normally associate with computer
science, especially in the research areas. So as humans, we actually generate a lot of
data about how we're feeling, what we're thinking. We have body language. We have the way that we
speak, kind of the faces that we make. It's really difficult to process all of that data all at once. So we need
the help from computers and technology to not only capture all of that information, but also help us
make sense of that data by analyzing the information. Computers have been ubiquitous in our lives,
and we expect more meaningful interactions with our technologies and we want our technologies to understand this in some sense.
And I think there's a huge opportunity, an unmet need of our technology being more understanding of us and us being more understanding of what the capabilities of the technology are so we can have this rich, meaningful experience collaborating with the technologies and working
with the technologies. So in that sense, making our technology more emotionally aware
is a key part in making that experience meaningful.
So there's a lot to unpack there in terms of how you translate human emotions into machine language. And we'll
get there in a second. But I want to go back to this idea of human-centered design. To me,
that implies there's been a significant amount of design that wasn't human-centered. What can
we do to change that? So I think we miss the mark in two ways. One is that we often simplify humans because it makes the problem more elegant and easy to solve.
And we see this happening in machine learning research.
You know, humans are mindless, label-generating oracles.
You know, if we collect enough labels on whether a person on this photo is happy or not, we can build emotionally intelligent models.
But the truth is humans are a lot more complex than
that. And we tend to ignore the things that are really difficult. So, you know, I talked to a lot
of data scientists about bias in machine learning. You know, everybody knows that bias is a real
problem, but because it's hard to tackle, it's not their area of expertise, you know, it's ignored
or deprioritized. So that's one thing. Yeah. Another thing is that, you know, we give ourselves too much credit.
You know, we assume that we know the problem and jump to solutions too quickly.
We say to ourselves that users don't know what they need if you ask them.
You know, I don't know if you know the famous quote by Henry Ford.
If you ask people what they needed, they'd say faster horse.
So I think if you do your user research right,
you'll not only find the right problems, but also through this iterative design process,
you'll discover the right solution that is the automobile. Not only that, you'll also be able to
contextualize your solution in more detail in real life situations. And, you know, the best part of my
job is actually talking to users about their problems and hearing their frustrations and
complaints. And really the hardest part is not jumping to solutions. And it takes effort and
discipline to stay true to your users' problems. And we often ignore this step. So talk to the
users. When we talked about human health and well-being,
we talked about some of the problems human practitioners face. And you mentioned two
specific ways that technology might help us attack these problems. And they were sensing
and intervention. Could you unpack these concepts for us and tell us how they might play out
in the human and technical worlds? How might technology augment and fill in human gaps here?
I want to use productivity as an example, since we are a productivity company.
Right.
So, you know, people go through a wide range of emotions throughout their workday.
And we have certain events or experiences, maybe a bad meeting, which could trigger negative emotion states like anxiety
and stress. And that leads to unhealthy behaviors or loss of productivity or even poor decision
making. So, you know, first step of coping with negative emotional states like stress would be
awareness, which not only involves understanding the triggers that led to the stress,
but reflecting how your stress is presented in your behavior. So that's when sensing comes in,
because it's difficult for people to associate these triggers with their behaviors and the
emotions. So multimodal sensing assists people with the collection of contextual data, and it might
actually help you learn salient features and patterns that actually correlate the stress
and negative emotional states to your behaviors, right?
So that's the multimodal sensing part.
And then there's the intervention part where we want to help people make sense of their
behaviors and actually do
something about it. Perhaps interact with the sensing technology to tweak it, personalize it,
you know, which is an important machine teaching moment, and work with this emotionally intelligent
agent to maybe surface interventions at the right moments. You know, so you can imagine having
adjusted time, cognitive reframing for some people. This could be after-the-fact reflection for some other people. Studies have shown that the best time to practice some of these interventions is when the person is at his or her best self. So providing the right intervention at the right time is key for sustaining engagement in mental health care.
Let me interrupt here. What might that look like, a just-in-time intervention? If I'm stressed and my device, whatever it is that senses it, alerts me and proposes that I reflect and so on,
do I have to wait till I'm at my better self so I can intervene on myself, as it were?
I think that's where the personalization comes in.
Everybody's different.
So imagine you have anger management problems.
I don't have to imagine.
And some intelligent thing tells you, hey, you're about to have an angry bout, you would be even more angry.
Or some people want to get interventions. Some people want others to tell them what to do.
We have done studies in the past where we've analyzed different types of people,
what kinds of control they want to have. So people who are more, you know, quote unquote,
control freaks are less likely to want to engage
in these just-in-time behaviors
and people who are less control freaks
are more likely.
So we really need to personalize
and really understand the person
and be able to give the control
to the person to adjust
and tweak these settings.
I want to talk about the collaboration between the University of Washington
and Microsoft Research.
I know you've got a foot in both places.
So we just talked about bridging gaps
between humans and machines
and how one can complement the other.
But there's also gaps that exist between people.
Talk about why it's important to have both medical
and technical experts working together
in this hybrid space.
We try to make meaningful contributions
to the space of human well-being
and hopefully make notable differences in mental health.
We really need expertise in clinical psychology.
But because we don't have the expertise in clinical psychology in-house, we have to work with people from outside of our company.
So our Pocket Skills, which is a mobile app that we developed that Mary Cerensky, our manager,
spearheaded. And we collaborated with Marsha Linen, who is the creator and expert in dialectical
behavior therapy, DBT, and Chelsea Wilkes, who is a clinical psychologist. We collaborated with
them to translate this proven in-session therapy technique to a digital format in a mobile app.
And currently, I'm working with University of Washington psychiatry and behavioral sciences
faculties to develop technologies that we're actually going to build and test its clinical
efficacy. So again, without the experts in these fields,
we are not able to truly validate the efficacies of these technologies. I think there is a
significant community within the medical field where they develop these apps and they actually
do test for clinical efficacy. But on the technology side, I think we have an opportunity
to inform the design and the implementation of some of the technologies in the medical field.
And we also have an opportunity to gain the knowledge and expertise that medical
specialists have and bring it to technologists so that we actually do build things that are useful.
Let's talk about a specific project you're working on called, are you ready?
Designing Technology-Enhanced Collaborative Care for Cancer Patients with Depression.
Usually long titles like that have some clever acronym, but no, this one doesn't even make a word.
I don't know. Do you have a nickname for it?
It's really hard. Someone's got to figure a name out that rolls off the tongue. In any case,
I want you to talk about this research in detail because our listeners may not even know much about
the topic, let alone the technical approach to helping with the intervention. So what are some
of the underlying problems in what you call collaborative care? And why do you think technology could help?
So I think I need to explain what collaborative care is first.
So a collaborative care model is one implementation model where basically behavioral health care or mental health care is tightly integrated into whatever medical care setting.
So it could be primary care. And in my case, it's cancer care. The idea is that, you know,
cancer patients do develop mental health disorders because they're going through a tremendous stress.
If the medical team discovers that you need this additional mental health support because you have depression, anxiety, and other
things. They engage what they call care coordinators or care managers to help you treat
your psychosocial needs. And you have psychiatrists if you need additional medication support,
or if you develop a mental health disorder that's complex, you involve psychiatrists
who can not only consult on the cases, but also provide kind of medication prescription
support.
So it's this idea of putting the patient at the center and having the oncologist or providers
or physicians work with your care providers who deliver the day-to-day behavioral
health care and the psychiatrist who's giving guidance on medication and prescriptions. So
that's collaborative care in a nutshell. And these care managers, they're often social workers.
And not only are they supposed to give you this behavioral health care, whether it's, you know,
BA, behavioral activation, or CBT, cognitive behavioral therapy or whatever these interventions and treatments.
Not only are they tasked to do that, they're also tasked to do these navigational support things, you know, financial aid, making sure you have housing, making sure you're taking your medication, making sure you can get to the cancer center and get the care and get the infusion and treatment and whatnot. So they're overwhelmed. They're truly overwhelmed. So
what are some ways that we can reduce the workload of these social workers?
So one of the things that we notice is that these social workers are having to do a lot of things
during their sessions with their patients, including things
like administrating validated mental health scales. So PHQ-9, which measures depression,
and GAT-7, which measures anxiety. And that takes away 10 minutes out of a 30-minute session.
So one-third of your session is spent on administrating this thing.
So, you know, you could do it outside of the session.
You could do it at the waiting room.
That's great.
But in reality, these patients come in and out of the clinics almost daily because they're getting infusions.
So they're really burnt out and they don't want to see another person.
They just want to go home. So, you know, adding
this additional burden, like you have to stay in the waiting room for this much longer to fill out
this form, or sometimes they cancel appointments and reschedule appointments because they're just
burnt out. What are ways to help these patients engage in mental health care on their own,
you know, sometimes without having to come to the clinic, and then also help social workers be less burdened with the workload that they have.
So one of the things that we could do is, you know, offline remote assessment.
So have technology that pings the patients once in a while and measures their depression and anxiety.
And when that scale goes up, you know, if you measure high depression
all of a sudden, bring them in for care. Is that self-reporting then? Okay. Yeah, most of these
scales are done as a self-reporting tool anyways. They get handed a piece of paper. But sometimes
it's in a supervised setting. And my question would be, if I'm depressed and I'm alone,
am I just going to hit the snooze button?
Right. And that already happens even without technology, right? A lot of the patients don't want to come into the clinic, especially when they're depressed, they don't want to engage. So
that's already a challenge. But most of the times these patients are sitting in their bed,
using their phones or watching TV, using these opportunities
to bring up, you know, hey, it's really important to know where you are in terms of your depression
and anxiety scale. Like, do you want to self-assess right now? Or, hey, it looks like you've been
sitting at your bed for a long time. You need to go out and take a walk let's go do that one of the things that we discovered that patients are missing in this traditional
mental health care sense is that because social work in cancer settings it's not billable which
means that there is no formal scheduling for them which means they don't get to go through the
standard process that other providers do. Basically, they don't get treated as an official
provider, which means there's no after-visit summary that you get when you go see your
primary physician, right? You get this little printed out paper that says, you're supposed to
do this, go take these medications. These social workers don't get to do that because they don't even have a scheduler.
So a lot of times what you end up discussing during the sessions are forgotten
or it's written on a sticky note and lost.
But we also know that people carry around their cell phones all the time.
So if they could have the after visit summaries and notes and their action items, things that they are supposed to do.
If we can have technology that nudges them in the right way to actually take those actions and also help them remember what they're supposed to do.
Because it's written down and it's recorded in your app.
That's a huge area of opportunity that is not there currently.
We're at the point in the podcast where I ask what could possibly go wrong. I'm imagining as we talk that I'm a cancer patient and depressed.
And anyone who's ever struggled with depression for any reason knows how difficult it is to do anything, even if you're reminded.
At what point, if I have my phone telling me something, is there a handoff where there's a human? Because I might just ignore my machine, but somebody who hugs me or somebody who's talking to me face to face and says, Gretchen, I'm here for you. We can do this together. Is there a bridge to the human through what you're talking about in your interventions?
Right. What we're trying to do is not replace humans in any way, but augment the human capabilities.
I think the social workers are already overloaded.
So if we can reduce their workload by making their work a lot more efficient, now they have the opportunity to give a call to their patients and say, hey, how are you doing? One of the things that we discovered is a lot of the patients
go through these crisis moments and they want to talk to somebody. They're not suicidal,
so they don't see the need to call the suicide hotline, but they're just going through a hard
time. And we asked them, you know, how can we help support that? And most of the time people said, I just want to text with my social worker.
I just want to text with somebody that I know who understands my situation.
So we're trying to brainstorm ideas around how do we make that possible and also understand that social workers are burdened and they have a life too.
They might need the app as well.
Crises don't happen between the nine to five, right? The business hours. So how do we make
that possible by balancing the needs of both individuals? It's a really tough problem,
but I think there's a way, you know, obviously reducing the workload of the social workers and
having them have more time to engage is one thing. And giving the patients opportunity and maybe a social group and support network to talk about their issues would be another way. Even as we've talked, you've triggered some thoughts in my mind. Number one is the privacy issues that are involved, especially when they deal with my mental health.
So that's one area I'd like you to talk about.
And the other one is the shift towards machines and devices to meet our needs rather than humans.
I know that I've just unloaded a bunch of thoughts.
Unload back on me.
I think, again, I want to stress that we're trying to take advantage of technology as a mediator to help connect people as opposed to replace the human to human connection. So the metric that I would probably optimize for is not how often
does this patient use my technology, but how often does this person engage in the care and actually
engage with the social worker? So, you know, what keeps me up at night is do I have the right
objectives? And am I actually meeting the
goals of that objective in terms of the privacy I think one thing that people assume is you have
the sensing technology it's creepy you know it's watching me all the time but you also have to
understand we're not forcing you to have this we want you to use this technology for yourself, for your own good. So if you're not
comfortable, don't do it. But if you see the need and if you see the benefit of this technology
in helping you remember what happened, helping you reflect, helping you get out of that tough
situation, then by all means, why wouldn't you use it? What I'm trying to get at
is that there are some points in your therapy session where your social worker or your therapist
will ask, you know, what happened? And you may realize that you can't remember, you can't recall.
That makes this session less efficient. I think there are ways to negotiate this boundary that you have of your comfort level.
And that negotiation has to happen by you realizing through the help of your social worker and therapist.
Where's the line that you want to draw and what's the need that you have?
To me, it all goes back to settings. You know, I'm going to pick the strength of intervention or sensing that I want.
Yeah. But you as a patient and you as a user sometimes don't know where that setting has to be.
And it's really the therapist or the social worker that helps you decide, okay, this data is really important because with this
data, I'm going to make this clinical adjustment. And that's a buy-in process. The social worker has
to convince that that data is important for clinical reasons. You know, that's a super good
point because there might be people that really, really need it and say, I don't need it. And
that's where you're looking at them going, yeah, you do.
Let's work on getting you to the point where you can embrace.
Who's your audience for this?
Is this technology really for the social worker and the therapist, not so much the oncologist or the primary care physician?
Right now, we're targeting just the social workers and the therapists and the care managers because they're the ones that know everything.
They're the go-to person for the patient.
They're the hub of all the information.
Okay.
It's story time here on the Microsoft Research Podcast.
We know about the work you're doing, but not much about you as a person, except the fact that you're doing a lot of work and you've had a lot of experiences already in your fairly short life. What's your background? How did
young Gina Su get involved in high tech and how did she end up at Microsoft Research?
I studied physics as an undergrad and astronomy. I did research in astrophysics.
I got involved in particle physics and I aspireired to be an experimental particle physicist I worked
at CERN for a summer so I went on to get my PhD in physics and felt like I was making kind of a
very abstract contribution which if anybody's studying physics like by all means go for it
but for me personally I felt like I needed to
do something that was directly impactful to people. So I started looking around for jobs.
And there are only two kinds of jobs that you could get as a physics graduate, which is
quantitative trading and finance, or some sort of a data science or, you know, computer science
and tech. Yeah. So I got an internship at Microsoft.
I fell in love with it.
And I just decided I'm going to pursue my career in coding.
I don't have a computer science background at all.
The only class that I took was in Intro to C.
And there I learned how to manipulate some memories and arrays and things like that.
But I never really knew how to code.
So I picked
all of that up at work. Somehow I started as an esthet. I started as a tester and then I moved my
way through becoming a developer. So all of that happened while I was working in Xbox. And this was
during the launch of Xbox One. And I don't know if you understand the pressure that you feel when you're not only shipping software, but you're also shipping brand new hardware.
We were working day and night over the weekends.
And I had a four-year-old at the time.
And I asked myself, what am I doing here?
I have a four-year-old at home.
So right after the Xbox One release, I had
decided to quit my job. I wanted to become a stay-at-home mom and dedicate myself to my child.
And I had an opportunity to give a talk on behalf of somebody at Grace Hopper. And I was like,
okay, I'm just going to go do it. And you know, I told
myself, I'm going to quit my job after this conference, I'm going to tell my manager.
At the time, the keynote speaker was Sheryl Sandberg. And I heard about Lean In. I didn't
read it. But I figured not only is she the keynote speakers, I was also invited to a breakfast with her.
So I better know this person.
So my husband put Lean In on the Kindle.
I started reading it from the airport.
I read it on the plane.
I read it at the hotel when I got there.
I stayed up all night reading the book, just crying my eyes out.
Crying my eyes out.
Why am I deciding to quit my job?
I can't do this. I can't add to that
number of women who gave up. So that gave me a lot of hope to, you know, pick myself back up,
get back on the field. And so that's when I started looking into other job opportunities.
I need to get out of what I was doing. I need to do something that's meaningful. I need to do something that has good work-life balance. But also, you know, something I can be passionate about. And I look towards Microsoft Research because somebody else told me, oh, yeah, there are developer positions at Microsoft Research. Really? I didn't know that. So I applied. I had great interviews. I fell in love with the
people that I was talking with, with the project, the idea of helping people, empowering people,
developing these technologies, studying people. I just fell in love with it all. And that's how
I got started in Microsoft Research. As we close, I always ask my guests to give us some parting
thoughts. And I usually ask it in terms of advice to our listeners, but I want to tweak the question
and ask it less as advice and more as what excites you most about where you're heading
with research? What technology wave do you want to catch and ride so right now there are a lot of changes that are
happening in mental health you know we're putting mental health as the number one epidemic people
are not afraid to talk about it you know we see the epidemic in teens are becoming more depressed
right now we're seeing changes in policy we're seeing changes to health care in terms of mental health so I think this is the
right moment to really get into the space of mental health and have that conversation we need
to have that conversation now so that it's not too late 10 years from now we have this emotionally
intelligent agent you know having to figure out,
like, what does it mean?
How does it impact our lives?
If we get there at the very beginning of it,
I think we will evolve
with the rest of the world.
We will evolve with the policy changes
that are happening,
the insurance changes that are happening,
the attitude changes
and stigma changes that are happening.
So I want to be part of that conversation now. Gina Su, thank you so much for joining us today. It's been delightful. Thank you for having
me. To learn more about Gina Su and the latest in technologies for human health and well-being,
visit microsoft.com slash research.