Microsoft Research Podcast - 093 - Data science and ML for human well-being with Jina Suh

Episode Date: October 9, 2019

Using technology to help us improve our health is nothing new: a quick web search returns hundreds of apps and devices claiming to help us get fit, quit smoking, master anxiety or just “find our cen...ter.” What is new is a serious cohort of researchers exploring how artificial emotional intelligence, or AEI, could help us understand ourselves better and, when used in concert with human caregivers, enhance our well-being. One of those researchers is Jina Suh, a former Xbox developer who got hooked on research and is now an RSDE in the Human Understanding and Empathy group at MSR, as well as a PhD student in computer science at the University of Washington. On today’s podcast, Jina shares her passion for creating technologies that promote emotional resilience and mental health; gives us an inside look at an innovative research collaboration that aims to improve collaborative care for cancer patients with depression; and tells us an emotional story of how, on the brink of quitting her job, she found inspiration to get back in the game and begin a new career in research for human well-being. https://www.microsoft.com/research

Transcript
Discussion (0)
Starting point is 00:00:00 There's a huge opportunity, an unmet need of our technology being more understanding of us and us being more understanding of what the capabilities of the technology are so we can have this rich, meaningful experience collaborating with the technologies and working with the technologies. So in that sense, making our technology more emotionally aware is a key part in making that experience meaningful. You're listening to the Microsoft Research Podcast, a show that brings you closer to the cutting edge of technology research and the scientists behind it. I'm your host, Gretchen Huizenga. Using technology to help us improve our health is nothing new. A quick web search returns hundreds of apps and devices claiming to help us get fit, quit smoking, master anxiety, or just find our center.
Starting point is 00:00:58 What is new is a serious cohort of researchers exploring how artificial emotional intelligence, or AEI, could help us understand ourselves better and, when used in concert with human caregivers, enhance our well-being. One of those researchers is Gina Su, a former Xbox developer who got hooked on research and is now an RSDE in the Human Understanding and Empathy Group at MSR, as well as a PhD student in computer science at the University of Washington. On today's podcast, Gina shares her passion for creating technologies that promote emotional resilience in mental health, gives us an inside look at an innovative research collaboration that aims to improve collaborative care for cancer patients with depression, and tells
Starting point is 00:01:41 us an emotional story of how, on the brink of quitting her job, she found inspiration to get back in the game and begin a new career in research for human well-being. That and much more on this episode of the Microsoft Research Podcast. Gina Su, welcome to the podcast. Thank you. You work in an interesting space, Gina. Your title is Research Software Design Engineer. You're currently working in a group called HU, or the Human Understanding and Empathy Group at MSR.
Starting point is 00:02:25 And you're beginning your doctoral studies in computer science at the University of Washington. In broad strokes, tell us what motivates you to do the work you do and what gets you up in the morning. What gets me up in the morning is actually an alarm clock. No, but really it's my daughter. She's really my inspiration. I get up to do what I set out to do because I want to make the world a better place for her. I want to show that, you know, being a working mom is hard, but it's really rewarding. And that learning never stops. So I want to show her that you can redefine yourself at any age, under any circumstances. So I do work in technology for human understanding and well-being. So I think in terms of mental health and human well-being, I think all of us at some point go through experiences where we feel like
Starting point is 00:03:13 we're not in control of our emotions or our lives. We go through very stressful moments. We go through very sad moments. I think there's a huge opportunity for us to develop technologies that helps us understand who we are, what are the feelings that we're having, and really help us reflect and be the better version of ourselves. Where is it that we would need technology to help us be better versions of ourselves or to overcome obstacles in our emotional and mental health? I think in a lot of ways, we have the capability to kind of look at what we've done and reflect on our past and figure out why we're feeling certain ways. But that takes effort and motivation and training and practice too. So a lot of people need professional help through therapies to develop the skills in order to reflect.
Starting point is 00:04:13 But I think technology, especially with multimodal sensing technologies, we have the opportunity to bring information to the person and help them look at their behaviors in ways that human memories cannot fathom. Right. In some ways, what I'm hearing you say is that if we can set a timer or an alarm for some physical thing we have to do, why couldn't we have technology prompt us on things that we have to do for our mental health. Right. And a lot of our mental health influences our productivity and our work and our thinking and our decision making. So I think it's absolutely necessary for us to schedule for ourselves to reflect and also practice the skills that we need to do to reflect and be mindful, just like we practice for work and for exercises. Well, we'll get to the specifics on that in a second, but let's go back upstream a little bit. You have a dev background, but you say now you only code when you need to.
Starting point is 00:05:18 When do you need to? Well, when I have an idea. Okay, cool. I was wondering about that. You've also done some work in machine learning and HCI, and you say you're a proud alumnus of Patrice Simard's machine teaching group. Tell us a little bit about your previous work and how it's informed the work that you do now. So I started as a developer in Xbox.
Starting point is 00:05:41 I worked in Xbox for seven plus years and i always admired the design team and loved working with them and you know trying to understand the rationale behind their design choices and i wanted to do more high-level thinking you know be closer to our customers and you know build things that are more meaningful and have genuine reasons why we're building them. So I stumbled across something called HCI, or human computer interaction, or HCD, human centered design. And at the same time, I stumbled across this machine teaching group, which claimed to be this focus on like studying the people interacting with machines. And, you know, when I started working in the group, I also started my part time master's program at University of Washington in human-centered design and engineering.
Starting point is 00:06:31 So at first I was just helping the group build their machine teaching tool. And then I got to help build prototypes for user studies and then help run the studies and eventually design studies using the knowledge that I was gaining from my master's and also through work. And I just fell in love with the science of it all and really observing people using the tools in the way that you hypothesize. So slowly and maturely, I started doing more independent research, forming my own projects and contributing to research. And I'm pretty much hooked now. I am a firm believer of the machine teaching principles, you know, this idea of empowering the human teacher. And I believe I can take those learnings and apply in other domains.
Starting point is 00:07:12 Right. And some exciting ones, too. I want to talk a bit more in depth about this human understanding and empathy or HUE group. Tell us in more detail what's going on in Hue and who's doing what right now. The mission of the Hue team is to really empower people by creating and inventing new technologies that promote emotional resilience and well-being. It's really grounded in the fact that emotions are fundamental to human interactions and they influence everything that we do, starting from learning, memory, decision-making, and all these other aspects of our lives. So, you know, how do we bring emotional intelligence to technology is kind of the core of our research. Tell me a little
Starting point is 00:07:56 bit more about this idea of emotional intelligence and emotional resilience, because this isn't something we normally associate with computer science, especially in the research areas. So as humans, we actually generate a lot of data about how we're feeling, what we're thinking. We have body language. We have the way that we speak, kind of the faces that we make. It's really difficult to process all of that data all at once. So we need the help from computers and technology to not only capture all of that information, but also help us make sense of that data by analyzing the information. Computers have been ubiquitous in our lives, and we expect more meaningful interactions with our technologies and we want our technologies to understand this in some sense.
Starting point is 00:08:46 And I think there's a huge opportunity, an unmet need of our technology being more understanding of us and us being more understanding of what the capabilities of the technology are so we can have this rich, meaningful experience collaborating with the technologies and working with the technologies. So in that sense, making our technology more emotionally aware is a key part in making that experience meaningful. So there's a lot to unpack there in terms of how you translate human emotions into machine language. And we'll get there in a second. But I want to go back to this idea of human-centered design. To me, that implies there's been a significant amount of design that wasn't human-centered. What can we do to change that? So I think we miss the mark in two ways. One is that we often simplify humans because it makes the problem more elegant and easy to solve. And we see this happening in machine learning research.
Starting point is 00:09:52 You know, humans are mindless, label-generating oracles. You know, if we collect enough labels on whether a person on this photo is happy or not, we can build emotionally intelligent models. But the truth is humans are a lot more complex than that. And we tend to ignore the things that are really difficult. So, you know, I talked to a lot of data scientists about bias in machine learning. You know, everybody knows that bias is a real problem, but because it's hard to tackle, it's not their area of expertise, you know, it's ignored or deprioritized. So that's one thing. Yeah. Another thing is that, you know, we give ourselves too much credit. You know, we assume that we know the problem and jump to solutions too quickly.
Starting point is 00:10:31 We say to ourselves that users don't know what they need if you ask them. You know, I don't know if you know the famous quote by Henry Ford. If you ask people what they needed, they'd say faster horse. So I think if you do your user research right, you'll not only find the right problems, but also through this iterative design process, you'll discover the right solution that is the automobile. Not only that, you'll also be able to contextualize your solution in more detail in real life situations. And, you know, the best part of my job is actually talking to users about their problems and hearing their frustrations and
Starting point is 00:11:09 complaints. And really the hardest part is not jumping to solutions. And it takes effort and discipline to stay true to your users' problems. And we often ignore this step. So talk to the users. When we talked about human health and well-being, we talked about some of the problems human practitioners face. And you mentioned two specific ways that technology might help us attack these problems. And they were sensing and intervention. Could you unpack these concepts for us and tell us how they might play out in the human and technical worlds? How might technology augment and fill in human gaps here? I want to use productivity as an example, since we are a productivity company.
Starting point is 00:11:52 Right. So, you know, people go through a wide range of emotions throughout their workday. And we have certain events or experiences, maybe a bad meeting, which could trigger negative emotion states like anxiety and stress. And that leads to unhealthy behaviors or loss of productivity or even poor decision making. So, you know, first step of coping with negative emotional states like stress would be awareness, which not only involves understanding the triggers that led to the stress, but reflecting how your stress is presented in your behavior. So that's when sensing comes in, because it's difficult for people to associate these triggers with their behaviors and the
Starting point is 00:12:37 emotions. So multimodal sensing assists people with the collection of contextual data, and it might actually help you learn salient features and patterns that actually correlate the stress and negative emotional states to your behaviors, right? So that's the multimodal sensing part. And then there's the intervention part where we want to help people make sense of their behaviors and actually do something about it. Perhaps interact with the sensing technology to tweak it, personalize it, you know, which is an important machine teaching moment, and work with this emotionally intelligent
Starting point is 00:13:16 agent to maybe surface interventions at the right moments. You know, so you can imagine having adjusted time, cognitive reframing for some people. This could be after-the-fact reflection for some other people. Studies have shown that the best time to practice some of these interventions is when the person is at his or her best self. So providing the right intervention at the right time is key for sustaining engagement in mental health care. Let me interrupt here. What might that look like, a just-in-time intervention? If I'm stressed and my device, whatever it is that senses it, alerts me and proposes that I reflect and so on, do I have to wait till I'm at my better self so I can intervene on myself, as it were? I think that's where the personalization comes in. Everybody's different. So imagine you have anger management problems. I don't have to imagine.
Starting point is 00:14:17 And some intelligent thing tells you, hey, you're about to have an angry bout, you would be even more angry. Or some people want to get interventions. Some people want others to tell them what to do. We have done studies in the past where we've analyzed different types of people, what kinds of control they want to have. So people who are more, you know, quote unquote, control freaks are less likely to want to engage in these just-in-time behaviors and people who are less control freaks are more likely.
Starting point is 00:14:52 So we really need to personalize and really understand the person and be able to give the control to the person to adjust and tweak these settings. I want to talk about the collaboration between the University of Washington and Microsoft Research. I know you've got a foot in both places.
Starting point is 00:15:24 So we just talked about bridging gaps between humans and machines and how one can complement the other. But there's also gaps that exist between people. Talk about why it's important to have both medical and technical experts working together in this hybrid space. We try to make meaningful contributions
Starting point is 00:15:42 to the space of human well-being and hopefully make notable differences in mental health. We really need expertise in clinical psychology. But because we don't have the expertise in clinical psychology in-house, we have to work with people from outside of our company. So our Pocket Skills, which is a mobile app that we developed that Mary Cerensky, our manager, spearheaded. And we collaborated with Marsha Linen, who is the creator and expert in dialectical behavior therapy, DBT, and Chelsea Wilkes, who is a clinical psychologist. We collaborated with them to translate this proven in-session therapy technique to a digital format in a mobile app.
Starting point is 00:16:28 And currently, I'm working with University of Washington psychiatry and behavioral sciences faculties to develop technologies that we're actually going to build and test its clinical efficacy. So again, without the experts in these fields, we are not able to truly validate the efficacies of these technologies. I think there is a significant community within the medical field where they develop these apps and they actually do test for clinical efficacy. But on the technology side, I think we have an opportunity to inform the design and the implementation of some of the technologies in the medical field. And we also have an opportunity to gain the knowledge and expertise that medical
Starting point is 00:17:18 specialists have and bring it to technologists so that we actually do build things that are useful. Let's talk about a specific project you're working on called, are you ready? Designing Technology-Enhanced Collaborative Care for Cancer Patients with Depression. Usually long titles like that have some clever acronym, but no, this one doesn't even make a word. I don't know. Do you have a nickname for it? It's really hard. Someone's got to figure a name out that rolls off the tongue. In any case, I want you to talk about this research in detail because our listeners may not even know much about the topic, let alone the technical approach to helping with the intervention. So what are some
Starting point is 00:18:01 of the underlying problems in what you call collaborative care? And why do you think technology could help? So I think I need to explain what collaborative care is first. So a collaborative care model is one implementation model where basically behavioral health care or mental health care is tightly integrated into whatever medical care setting. So it could be primary care. And in my case, it's cancer care. The idea is that, you know, cancer patients do develop mental health disorders because they're going through a tremendous stress. If the medical team discovers that you need this additional mental health support because you have depression, anxiety, and other things. They engage what they call care coordinators or care managers to help you treat your psychosocial needs. And you have psychiatrists if you need additional medication support,
Starting point is 00:19:00 or if you develop a mental health disorder that's complex, you involve psychiatrists who can not only consult on the cases, but also provide kind of medication prescription support. So it's this idea of putting the patient at the center and having the oncologist or providers or physicians work with your care providers who deliver the day-to-day behavioral health care and the psychiatrist who's giving guidance on medication and prescriptions. So that's collaborative care in a nutshell. And these care managers, they're often social workers. And not only are they supposed to give you this behavioral health care, whether it's, you know,
Starting point is 00:19:43 BA, behavioral activation, or CBT, cognitive behavioral therapy or whatever these interventions and treatments. Not only are they tasked to do that, they're also tasked to do these navigational support things, you know, financial aid, making sure you have housing, making sure you're taking your medication, making sure you can get to the cancer center and get the care and get the infusion and treatment and whatnot. So they're overwhelmed. They're truly overwhelmed. So what are some ways that we can reduce the workload of these social workers? So one of the things that we notice is that these social workers are having to do a lot of things during their sessions with their patients, including things like administrating validated mental health scales. So PHQ-9, which measures depression, and GAT-7, which measures anxiety. And that takes away 10 minutes out of a 30-minute session. So one-third of your session is spent on administrating this thing.
Starting point is 00:20:45 So, you know, you could do it outside of the session. You could do it at the waiting room. That's great. But in reality, these patients come in and out of the clinics almost daily because they're getting infusions. So they're really burnt out and they don't want to see another person. They just want to go home. So, you know, adding this additional burden, like you have to stay in the waiting room for this much longer to fill out this form, or sometimes they cancel appointments and reschedule appointments because they're just
Starting point is 00:21:14 burnt out. What are ways to help these patients engage in mental health care on their own, you know, sometimes without having to come to the clinic, and then also help social workers be less burdened with the workload that they have. So one of the things that we could do is, you know, offline remote assessment. So have technology that pings the patients once in a while and measures their depression and anxiety. And when that scale goes up, you know, if you measure high depression all of a sudden, bring them in for care. Is that self-reporting then? Okay. Yeah, most of these scales are done as a self-reporting tool anyways. They get handed a piece of paper. But sometimes it's in a supervised setting. And my question would be, if I'm depressed and I'm alone,
Starting point is 00:22:03 am I just going to hit the snooze button? Right. And that already happens even without technology, right? A lot of the patients don't want to come into the clinic, especially when they're depressed, they don't want to engage. So that's already a challenge. But most of the times these patients are sitting in their bed, using their phones or watching TV, using these opportunities to bring up, you know, hey, it's really important to know where you are in terms of your depression and anxiety scale. Like, do you want to self-assess right now? Or, hey, it looks like you've been sitting at your bed for a long time. You need to go out and take a walk let's go do that one of the things that we discovered that patients are missing in this traditional mental health care sense is that because social work in cancer settings it's not billable which
Starting point is 00:22:59 means that there is no formal scheduling for them which means they don't get to go through the standard process that other providers do. Basically, they don't get treated as an official provider, which means there's no after-visit summary that you get when you go see your primary physician, right? You get this little printed out paper that says, you're supposed to do this, go take these medications. These social workers don't get to do that because they don't even have a scheduler. So a lot of times what you end up discussing during the sessions are forgotten or it's written on a sticky note and lost. But we also know that people carry around their cell phones all the time.
Starting point is 00:23:41 So if they could have the after visit summaries and notes and their action items, things that they are supposed to do. If we can have technology that nudges them in the right way to actually take those actions and also help them remember what they're supposed to do. Because it's written down and it's recorded in your app. That's a huge area of opportunity that is not there currently. We're at the point in the podcast where I ask what could possibly go wrong. I'm imagining as we talk that I'm a cancer patient and depressed. And anyone who's ever struggled with depression for any reason knows how difficult it is to do anything, even if you're reminded. At what point, if I have my phone telling me something, is there a handoff where there's a human? Because I might just ignore my machine, but somebody who hugs me or somebody who's talking to me face to face and says, Gretchen, I'm here for you. We can do this together. Is there a bridge to the human through what you're talking about in your interventions? Right. What we're trying to do is not replace humans in any way, but augment the human capabilities.
Starting point is 00:25:16 I think the social workers are already overloaded. So if we can reduce their workload by making their work a lot more efficient, now they have the opportunity to give a call to their patients and say, hey, how are you doing? One of the things that we discovered is a lot of the patients go through these crisis moments and they want to talk to somebody. They're not suicidal, so they don't see the need to call the suicide hotline, but they're just going through a hard time. And we asked them, you know, how can we help support that? And most of the time people said, I just want to text with my social worker. I just want to text with somebody that I know who understands my situation. So we're trying to brainstorm ideas around how do we make that possible and also understand that social workers are burdened and they have a life too. They might need the app as well.
Starting point is 00:26:07 Crises don't happen between the nine to five, right? The business hours. So how do we make that possible by balancing the needs of both individuals? It's a really tough problem, but I think there's a way, you know, obviously reducing the workload of the social workers and having them have more time to engage is one thing. And giving the patients opportunity and maybe a social group and support network to talk about their issues would be another way. Even as we've talked, you've triggered some thoughts in my mind. Number one is the privacy issues that are involved, especially when they deal with my mental health. So that's one area I'd like you to talk about. And the other one is the shift towards machines and devices to meet our needs rather than humans. I know that I've just unloaded a bunch of thoughts. Unload back on me.
Starting point is 00:27:06 I think, again, I want to stress that we're trying to take advantage of technology as a mediator to help connect people as opposed to replace the human to human connection. So the metric that I would probably optimize for is not how often does this patient use my technology, but how often does this person engage in the care and actually engage with the social worker? So, you know, what keeps me up at night is do I have the right objectives? And am I actually meeting the goals of that objective in terms of the privacy I think one thing that people assume is you have the sensing technology it's creepy you know it's watching me all the time but you also have to understand we're not forcing you to have this we want you to use this technology for yourself, for your own good. So if you're not comfortable, don't do it. But if you see the need and if you see the benefit of this technology
Starting point is 00:28:13 in helping you remember what happened, helping you reflect, helping you get out of that tough situation, then by all means, why wouldn't you use it? What I'm trying to get at is that there are some points in your therapy session where your social worker or your therapist will ask, you know, what happened? And you may realize that you can't remember, you can't recall. That makes this session less efficient. I think there are ways to negotiate this boundary that you have of your comfort level. And that negotiation has to happen by you realizing through the help of your social worker and therapist. Where's the line that you want to draw and what's the need that you have? To me, it all goes back to settings. You know, I'm going to pick the strength of intervention or sensing that I want.
Starting point is 00:29:12 Yeah. But you as a patient and you as a user sometimes don't know where that setting has to be. And it's really the therapist or the social worker that helps you decide, okay, this data is really important because with this data, I'm going to make this clinical adjustment. And that's a buy-in process. The social worker has to convince that that data is important for clinical reasons. You know, that's a super good point because there might be people that really, really need it and say, I don't need it. And that's where you're looking at them going, yeah, you do. Let's work on getting you to the point where you can embrace. Who's your audience for this?
Starting point is 00:29:51 Is this technology really for the social worker and the therapist, not so much the oncologist or the primary care physician? Right now, we're targeting just the social workers and the therapists and the care managers because they're the ones that know everything. They're the go-to person for the patient. They're the hub of all the information. Okay. It's story time here on the Microsoft Research Podcast. We know about the work you're doing, but not much about you as a person, except the fact that you're doing a lot of work and you've had a lot of experiences already in your fairly short life. What's your background? How did young Gina Su get involved in high tech and how did she end up at Microsoft Research?
Starting point is 00:30:34 I studied physics as an undergrad and astronomy. I did research in astrophysics. I got involved in particle physics and I aspireired to be an experimental particle physicist I worked at CERN for a summer so I went on to get my PhD in physics and felt like I was making kind of a very abstract contribution which if anybody's studying physics like by all means go for it but for me personally I felt like I needed to do something that was directly impactful to people. So I started looking around for jobs. And there are only two kinds of jobs that you could get as a physics graduate, which is quantitative trading and finance, or some sort of a data science or, you know, computer science
Starting point is 00:31:23 and tech. Yeah. So I got an internship at Microsoft. I fell in love with it. And I just decided I'm going to pursue my career in coding. I don't have a computer science background at all. The only class that I took was in Intro to C. And there I learned how to manipulate some memories and arrays and things like that. But I never really knew how to code. So I picked
Starting point is 00:31:45 all of that up at work. Somehow I started as an esthet. I started as a tester and then I moved my way through becoming a developer. So all of that happened while I was working in Xbox. And this was during the launch of Xbox One. And I don't know if you understand the pressure that you feel when you're not only shipping software, but you're also shipping brand new hardware. We were working day and night over the weekends. And I had a four-year-old at the time. And I asked myself, what am I doing here? I have a four-year-old at home. So right after the Xbox One release, I had
Starting point is 00:32:26 decided to quit my job. I wanted to become a stay-at-home mom and dedicate myself to my child. And I had an opportunity to give a talk on behalf of somebody at Grace Hopper. And I was like, okay, I'm just going to go do it. And you know, I told myself, I'm going to quit my job after this conference, I'm going to tell my manager. At the time, the keynote speaker was Sheryl Sandberg. And I heard about Lean In. I didn't read it. But I figured not only is she the keynote speakers, I was also invited to a breakfast with her. So I better know this person. So my husband put Lean In on the Kindle.
Starting point is 00:33:10 I started reading it from the airport. I read it on the plane. I read it at the hotel when I got there. I stayed up all night reading the book, just crying my eyes out. Crying my eyes out. Why am I deciding to quit my job? I can't do this. I can't add to that number of women who gave up. So that gave me a lot of hope to, you know, pick myself back up,
Starting point is 00:33:37 get back on the field. And so that's when I started looking into other job opportunities. I need to get out of what I was doing. I need to do something that's meaningful. I need to do something that has good work-life balance. But also, you know, something I can be passionate about. And I look towards Microsoft Research because somebody else told me, oh, yeah, there are developer positions at Microsoft Research. Really? I didn't know that. So I applied. I had great interviews. I fell in love with the people that I was talking with, with the project, the idea of helping people, empowering people, developing these technologies, studying people. I just fell in love with it all. And that's how I got started in Microsoft Research. As we close, I always ask my guests to give us some parting thoughts. And I usually ask it in terms of advice to our listeners, but I want to tweak the question and ask it less as advice and more as what excites you most about where you're heading with research? What technology wave do you want to catch and ride so right now there are a lot of changes that are
Starting point is 00:34:46 happening in mental health you know we're putting mental health as the number one epidemic people are not afraid to talk about it you know we see the epidemic in teens are becoming more depressed right now we're seeing changes in policy we're seeing changes to health care in terms of mental health so I think this is the right moment to really get into the space of mental health and have that conversation we need to have that conversation now so that it's not too late 10 years from now we have this emotionally intelligent agent you know having to figure out, like, what does it mean? How does it impact our lives?
Starting point is 00:35:28 If we get there at the very beginning of it, I think we will evolve with the rest of the world. We will evolve with the policy changes that are happening, the insurance changes that are happening, the attitude changes and stigma changes that are happening.
Starting point is 00:35:43 So I want to be part of that conversation now. Gina Su, thank you so much for joining us today. It's been delightful. Thank you for having me. To learn more about Gina Su and the latest in technologies for human health and well-being, visit microsoft.com slash research.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.