In The Arena by TechArena - AI Assisted Therapy with Lyssn's Zac Imel

Episode Date: November 23, 2022

TechArena host Allyson Klein talks with Lyssn co-founder Zac Imel about how his company intends to change the shape of mental health using artificial intelligence....

Transcript
Discussion (0)
Starting point is 00:00:00 Welcome to the Tech Arena, featuring authentic discussions between tech's leading innovators and our host, Allison Klein. Now, let's step into the arena. Welcome to the Tech Arena. My name is Alison Klein, and today I'm delighted to be joined by Zach Immel of Lyssen. Zach, welcome to the program. Hi, thanks for having me. So Zach, I had you on today to talk about how technology is intercepting with the mental health arena. Why don't you give me a little bit of background on you and of Listen, the company that you helped found a few years ago? Absolutely. Sure. But I guess in terms of my background, I'm a psychologist by training and have a PhD in counseling psychology and worked as a therapist in the VA for just a little bit
Starting point is 00:01:04 and then have been a professor at the University of Utah and director of clinical training for maybe 10 or 11 years. And really, ever since I got started in psychology, I've had kind of an interest in data science, what we call data science now used to be, you know, statistics. And so I've always had a bit of an analytical mind. And the way I got interested in psychology was, how can we apply whatever the kind of current, most advanced statistical applications are to the questions that are really interesting to people in psychology. And for me, my primary interest was in psychotherapy research. So how and why does psychotherapy work, which is the conversation that a therapist has
Starting point is 00:01:51 with their patient. I was fascinated when I learned about the story of LISN and what you have put together. Most people think about psychotherapy as something that is distinctly not technical, a conversation over time between two people and something where technology really doesn't enter the arena. What was the impetus for looking at how technology could apply and where were the areas of treatment that you were looking at that you could improve or help practitioners improve in terms of the way they were treating mental health? Yeah. I mean, for a long time, they would have been right, right? I mean, psychotherapy is the kind of with a goal in order to help change a behavior or change someone's emotion, we've been doing some form of psychotherapy probably forever, right?
Starting point is 00:02:51 I mean, we didn't call it that and didn't have a formal profession around it. And so, you know, really when I first started training in psychotherapy, if you asked about technology, that would have been, you know, we probably just started having CD-ROMs that where you could get like cognitive behavioral therapy and you could, you know, click on some buttons and read some stuff on a website. And, and, or it was like a trainee might have a bug in their ear where they're in a room talking to a client, but there's a supervisor on this other one-way mirror and they're listening in. And if you get stuck, the supervisor's kind of like, say this, right? Or do that or try this or that sort of thing.
Starting point is 00:03:29 And so that was technology really for quite a long time. And so the primary thing we identified, and this is going back, I think the first work we started doing with my colleague, Dave Atkins, and a bunch of folks in computer science and electrical engineering was the way we evaluate quality in psychotherapy is for a human being to listen to the conversation. And that's great. It's a nice way to do it. And we've been doing that for a long time. We have measures for that. They're reliable. We can tell you, some people find this surprising. In our field, you wouldn't. There are reliable measures for how empathic a therapist is, right? So I can train someone to listen to a conversation and, you know, we can score the person who's doing the interview, someone like you, for example, about like,
Starting point is 00:04:20 how well are they trying to actively understand the person who's talking to them are they actively asking for clarifications asking open questions being supportive or are they kind of barely paying attention right and certainly at the extremes we can capture that really well it just takes a long time and so that's never used outside of like really well-funded clinical research that's funded by like nih and you do that for ensuring internal validity of your interventions and making sure people are doing what you hope they're doing. So when you do a clinical trial, you know, oh, well, this worked because people did the intervention we thought they were doing. Psychotherapy is really different than, you know, pharmacotherapy, drug therapy, where, you know, you give someone a pill and factory has hopefully ensured that the thing that you think is supposed to be in the pill is in it.
Starting point is 00:05:08 And psychotherapy is a lot more like teaching, right? You synthesize the specific ingredients of psychotherapy in the moment with the person. They aren't there until you have the talk. And so the question is, how can we evaluate that in a way that's scalable at all and reliable? This was right around when speech and language technologies were starting to move into a new domain. I mean, we didn't have transformer models yet or some of the stuff that's come out in the last few years in machine learning. But, you know, speech recognition had gotten a lot better. And so we had the idea that what if we got some of our engineering colleagues who were working on these kind of
Starting point is 00:05:51 standard language corpus areas that were pretty well mined and got them interested in some of the problems we had in mental health care. Like we're having these really high stakes conversations. We don't have any way to analyze them at scale and we're having millions of them. Could we try and train some models that could at least start to replicate human evaluation of those things? So I believe the beginning was looking at ways that you could use this to train psychotherapists and give them feedback on their sessions. Is that correct? Yeah. I mean, it's a little of both. In some ways, training and ongoing quality are somewhat linked. And so I would say the first stuff we did was all like, could I listen to, if I had a 500 recordings from substance use counselors all over the Pacific Northwest and California. And I had human ratings of those. Could we replicate those human ratings? And so a lot of our early papers were just that,
Starting point is 00:06:53 you know, they were basic proof of concept. Can we do this? And there was a lot of skepticism that we could, especially in the clinical domain, Less so, I think, in the computer science domain. But after that, the first clinical application was more, okay, well, we have these models that we think kind of work. We build a platform and take it with folks who don't know much about a particular intervention. And maybe they've had an initial training and they can record a session, get immediate feedback on what they're doing. So instead of what you typically would get in a training is you would, this isn't the best case scenario, but it's more typical scenario. You would record like a little role play with a fake patient, maybe another colleague, and you'd record that and you get a tape, you know, or like a flash drive and you'd like mail it to someone. You know, maybe if you're really advanced, you'd put it on Dropbox.
Starting point is 00:07:47 And then you'd get feedback on that like in a month or in a couple of months. And, you know, we our first study we did on this, we give it to them in minutes. Right. And so there's a lot of like learning theory on why that's much better. Right. I mean, feedback should be proximal to the behavior you're trying to shape. And so we were able to vastly speed up that process and start to give people feedback much more quickly.
Starting point is 00:08:11 And from that, you've developed an entire suite of tools for this space for practitioners. Can you talk about the full suite and how that's evolved over time? Yeah, so at this point, you know, at some point we launched a company out of our university-based research, which you mentioned, and we've built really three lines of products. And so one of them, which we kind of call like always on quality monitoring, which is
Starting point is 00:08:39 that for digital mental health care, other spaces, we can just be on in the background. Of course, people are aware and consent to these things, but we can take in the conversations you're having, process them pretty much automatically and have that running all the time. And so you don't have to have people who are using and usually these are well-trained people that you'd rather actually spend, have their time being used to see patients. And so we have a dashboard and places where you can go and log in and look at your metrics, both as an individual practitioner, you can log in and see what happened in your session right away and how your sessions went and get some summative feedback across not just one session. So it's not just like one example of how you did, but across 10, 20, 50 sessions. Then your supervisor also, or administrators can do the same thing across hundreds or thousands of sessions. Then we've just now started to release some documentation
Starting point is 00:09:37 assistance tools. And so one of the big things that providers rightly complain about, it's probably one of the reasons I'm not a therapist anymore, is documentation is no fun. You get into the field as a person who's interested in people, and a lot of what you end up doing is treating the computer by writing notes and clicking boxes and all that sort of stuff. And so we've started to build some things where we can capture the content of the conversation and use some of our models to summarize it automatically in the style of a clinical note that at least gets it drafted for you. So you can look at it, edit it, and then speed up some of that process, maybe jog your memory of the session. Then we have primary training tools. And so where we take some people who've had no background in intervention at all and take them, give them some basic introduction into cognitive behavioral therapy, for example, which is a well-known intervention in our field and teach them a little bit about it.
Starting point is 00:10:36 And then instead of just like having you stare at slides, which is what you do mostly in online trainings, you just start practicing. We give you a little vignette of a case and you respond to it. And then we can score it using our machine learning models right away, give you that feedback. You can use that to shape your behavior a little bit, and you can do things repeatedly and show some skill growth pretty quickly. And so we have a couple of different products along those lines as well. Take us under the covers a little bit in terms of the underlying AI that you're utilizing to drive this model. I know you have something like 50, I don't remember if it was 56 or 57 parameters that you're scoring against. Can you talk a little bit about that and how you chose those?
Starting point is 00:11:22 Yeah. So I think what you're referencing is the different analytics we generate. Yeah. So we have, I think the last time, it's probably more than this now, but it's 54 different unique analytics that are broadly related to behaviors that the therapist is using or things that were discussed during the session. To me, there's kind of broad classes of things I've seen in the field and then in digital mental health now where, you know, you use a machine learning model to pick out words that you think are associated with a particular intervention. And so one example of that might be homework.
Starting point is 00:12:01 And in cognitive behavioral therapy, a big part of the intervention is that you're supposed to do homework. The therapist assigns stuff for you to do once you go home. And that's important part of extending the treatment hour into the client's lives. And a very, very simplistic version of that would be just to say, did the therapist say the word homework, right? Which might be more of an N-gram type of model and where we're just looking at specific keywords. That's fine. That's probably better than nothing. It doesn't map very well onto the human rated expert defined category of did the therapist actually do a good job talking about this particular thing with the client. And so what we've done is, and this goes
Starting point is 00:12:47 back to our university-based research and now is we pick gold standard measures from the field, right? And so there are existing measures in cognitive behavioral therapy or in an intervention we, it's called motivational interviewing, which is a substance use disorder treatment that we we have metrics for. We pick those measures and then we have a human coding team. That's all, you know, people with backgrounds in the field, licensed clinicians, things like that. They rate sessions. We assess internal inter-rater reliability. And we do it long enough until eventually the machine can replicate those human ratings at some level of agreement with the human raters. And so we assess percentage of human agreements and typically target like 80% of human agreements. And for a lot of things,
Starting point is 00:13:38 we beat that and are pretty much indistinguishable from human raters. And so the analytics we chose are based on what are the most widely researched and evidence-based practices in the field. So motivational interviewing and cognitive behavioral therapy are the most standard, well-researched, evidence-based interventions in mental health care. And so we started there. We thought we'd pick the ones that were most well-studied and use those tools. And so really what we've iterated on is not so much the, the, the human ratings that go into models, but the models themselves. And so we started doing this work on the university side back, you know, 2008, where if, if your listeners will know what this is, where we're using things like topic models
Starting point is 00:14:26 and other things where now we're using kind of the most cutting edge transformer based models, out of GBT world and Roberta and things like that, where there's models that are pre-trained on large corpora outside of our domain, but then we're doing extensive secondary training on our own internal data sets to make sure that we're really tracking the things we want to be tracking. So you're further honing four-year unique scenarios and then driving a tremendous amount of inference against clinician data. Is that correct? That's right. Yeah. So our gold standard for at least the things I was just talking about is human perception of those things. Right. And so if to just make it more concrete, like in motivational interviewing, one of the things you want therapists to do is make listening statements.
Starting point is 00:15:16 And so literally what you just did to me. Right. I said a long thing. And then you kind of came back and said, well, so it's this. And I went, yeah, that's right. And so that, that little intervention, we would call a reflection or a restatement in interviewing skills. And it's our key therapist behavior, right? If you can't do that, you're going to really struggle to be a therapist. And, and so we have humans who rate those things repeatedly thousands and thousands and thousands of times. Right. And then at some point we, we evaluate whether or not our human, our machines can rate those just as well. Fantastic. Now, if I was a therapist listening to this, I'd be like, first of all, fantastic.
Starting point is 00:15:57 You're going to help me and you're going to help me with my notes. All the therapists that I talk to say that that's the pain point of the profession. So that's the pain point of the profession. So that's awesome. But then there's a little voice inside my head wondering, Zach, is one of your ambitions to replace the human to human contact in therapy and maybe replace it with machine to human interactions? Can you talk a little bit about that? It's so funny that that's, I mean, I guess it's not funny. It makes sense that people wonder about that given what's out there in the world right now. If I were a therapist, and I guess I am sort of still, I would not be worried about robots replacing me anytime soon. I mean, maybe for folks in grade school, I don't know. Things
Starting point is 00:16:41 change. We've solved problems faster in some domains than we thought we would. And there's a lot of bot-based therapy that's out there now. And I don't know how much you've played around with some of those or talked to people who have, but I think they're really cool. And I play with them all the time. They are nowhere near replacing a human conversation. And I think some of that is the reliability of the bot is almost inversely related to the interestingness of the AI that's in it because these new generative models are really cool. And they're really, you know, almost creepily good at being able to replicate human conversation and say human-like things. But they're also, they can say things that you don't expect and they can be unreliable and they
Starting point is 00:17:32 can have all sorts of unintended behaviors. And so, you know, that's almost never what you're seeing in these bots that are out there in the world. I won't name any in particular. You're seeing stuff for the most part that was written by humans, therapists. And then there's some algorithm in the background that might be picking up, oh, they said the word anxiety, so I should say this versus that. But that's been pre-written, or it's kind of pushing you into a particular worksheet that's already pre-written or pre-designed or something like that. I think the future of mental health care, at least in the near term, is much more about augmenting therapists rather than replacing them. How can we make the job of
Starting point is 00:18:14 the humans who are doing the work easier? How can we make them more effective at their work, help them be less burned out, support them when they're struggling because they're doing so much work? And you know, it's one thing to be empathic for an hour. It's another thing to be empathic for 30 hours and then for 50 weeks. And so how do you support someone who's doing that? And I think that's to me where the interesting technology is. And it also seems like that's been the history of technology. It generally hasn't been, oh, we found this thing. Humans are gone now. Right.
Starting point is 00:18:49 Thinking about it, you're training algorithms based on human experience and you're augmenting based on that. And where I see it going is, are there other tools that we can provide that perhaps address some of the high level issues with people before they actually get into a therapist appointment. One of the things that I've been thinking about is just the lack of mental health professionals versus the demand in communities. And are there other things that we can do to help those folks that are struggling to find mental health care? Yeah. And I think what you're seeing is a lot of places are adding either bots or apps as a part of the kind of broad menu of services they provide. Right. And so it's, it's, I don't
Starting point is 00:19:38 necessarily think it really has to be an either or right where we certainly don't have enough therapists at all. And so there's a lot of people that really the question isn't, oh, wouldn't it be nice if they had a therapist, they don't have anything. And so can we give them something that's better than nothing? And I think the answer is, there's been some good studies that show we can give them some kind of brief interventions, particularly with people who are lower in acuity or severity, who aren't actively at risk for high risky behaviors, things like that. But for someone in my domain, whose background has been studying the conversation that therapists have with patients, when I typically get asked that
Starting point is 00:20:18 question, people are often asking me, do you think a robot could replace that? And I just don't think we're anywhere close to that in a way that could be scalable. What I do think could be interesting and stuff we're working on is rather than just evaluating, was this session empathic? Did they do the things they were supposed to? In particular cases, can we nudge people in different directions at the moment they're engaging with the client? So if we know we want people to be in certain circumstances, more actively empathic and expressing how hard they're trying to understand the client, can we suggest things they might say
Starting point is 00:20:55 before they say them? And I think the answer there is we could do that, right? Especially if there's a human still in the loop where who's making decisions about, well, no, I don't want to say that. These are just options. Yeah. So basically you're looking at predictive prompts based on the analysis of what's being said. That's interesting. Now, I know that the good news for Listen is that there's been incredible uptick and interest and deployment of this technology, including statewide in Utah. Can you tell me a little bit about where you've seen success and what's coming next as we head into the new year? We have had a bunch of success.
Starting point is 00:21:36 I think we're really what we're seeing is, you know, over the pandemic, you saw the beginning of like I've heard this mentioned, like kind of telehealth 1.0, which was basically saying, now can we do some psychotherapy over digital medium? And people are starting to try and think about like, well, what's next in terms of augmenting this beyond just making it easier for people to access care? They don't have to go sit in an office. And so I think there are, there are starting to be enough calls in the field for how do we scale quality, not just access? How do we make sure not just people are getting anything, but they're getting something that's of decent quality?
Starting point is 00:22:15 And that in particular has been found in large kind of public provider contracts where folks have to provide some evidence that they're doing stuff at a certain reasonable standard of care. And once the legislative requirement is there for people to do that, the burden on the providers is pretty significant to try and meet that. Because it's back to the stuff we were talking about before, humans rating this stuff when they could be doing other things and when it's hard to hire people. And so I know from being a professor, when you ask start people to do tasks repeatedly for hours, for years, they tend to have other opinions, right? And they eventually go do something else. And so we've had some success signing contracts with state
Starting point is 00:23:04 governments where we're helping them meet those requirements and not replacing humans out of the evaluative loop entirely, but just massively augmenting them. And so we now have contracts with the state of Utah, with the state of Wyoming, with Washington, D.C., with, I think, a county in Minnesota, a few others. And so then we've also just had good traction with traditional mental health clinics that are trying to scale up quality and figure out how to do better for their therapists who are getting burned out, but also want to provide better care. The training solutions that we're offering are much more compelling than what's typically out there. So if you're trying to do post-graduate education for your providers, which is almost always required, you've got to be doing something to maintain your license. And often what that looks like these days is some sort of online education
Starting point is 00:23:55 in a particular intervention. There's just not a lot of skill building that can happen in those places. And so we can really help people practice and get better at things. And so we've had a lot of organizations are pretty interested in that. So I mean, in the new year, I would expect we're going to end up with five or six, if not more, pretty soon, state-level contracts. We're starting to reach out to another kind of larger community-based mental health clinics, those sorts of places as well. I just want to commend you. I think this is such an interesting use of artificial intelligence to make an actually incredible impact on society. So I can't wait to hear more about what you do in the future. One final question for you, Zach, and thank you for your time today. Where can folks
Starting point is 00:24:40 go to find out more about what you and the LISN team are doing and engage with you guys? Our website is L-Y-S-S-N dot I-O, LISN dot I-O. We have links to research papers there, case studies, kind of a blog talking about the different things we're doing. And so if you want to reach out to us there, please do. Thanks so much for coming on the program today. Thanks for joining the Tech Arena. Subscribe and engage at our website, thetecharena.net. All content is copyright by the Tech Arena.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.