ACM ByteCast - Kristian Lum - Episode 4

Episode Date: August 6, 2020

In this episode of ACM ByteCast, Rashmi Mohan welcomes Kristian Lum to the podcast. Lum is part of the research faculty at the University of Pennsylvania's CIS Department. Previously, she was Lead Sta...tistician at the Human Rights Data Analysis Group (HRDAG), where she led the project on criminal justice in the United States. She's widely known for her work on algorithmic fairness and predictive policing and is a key organizer of the ACM Conference on Fairness, Accountability, and Transparency (ACM FAccT). Lum discusses her transition from math and statistics into computer science, and how her lab work in bioinformatics expanded her interest into social issues. They touch on the sensitive nature of data and privacy and gray areas in criminal justice data collection. Lum mentions some applications of her work, advising the NYC Mayor's Office of Criminal Justice and partnering with the ACLU. She also describes a very timely project accounting for the time lag between COVID-19 infection and death. Finally, she traces her winding, fascinating career path from academic to industry and back.

Transcript
Discussion (0)
Starting point is 00:00:00 This is ACM ByteCast, a podcast series from the Association for Computing Machinery, the world's largest educational and scientific computing society. We talk to researchers, practitioners, and innovators who are at the intersection of computing research and practice. They share their experiences, the lessons they've learned, and their own visions for the future of computing. I am your host, Rashmi Mohan. The use of technology for social good is something we all dream of.
Starting point is 00:00:32 Our guest today has lived that dream for many years. Christiane Lum has led the Human Rights Data Analysis Group as lead statistician, focusing her efforts on the uses of machine learning within the criminal justice system. Christiane, welcome to ACM ByteCast. Thanks so much for having me. I'd like to lead with a question that I ask all my guests. Please introduce yourself and talk about what you currently do and give us some insight
Starting point is 00:00:58 into what drew you into the field of computer science. Yeah, so actually, you're right that I have been the lead statistician for the Human Rights Data Analysis Group for quite some time. But as of about a month ago, I joined the research faculty at the University of Pennsylvania in their computer and information science department. So that's kind of an exciting update to my affiliation. So what got me into computer science? You know, I started out in math and statistics, both, and there really is quite a bit of overlap between statistics and machine learning. And so I don't really even necessarily see there being these really stark disciplinary boundaries. This sort of just seemed like a
Starting point is 00:01:36 natural transition based on the types of things I was working on and the sorts of affiliations that other people had in computer science relative to statistics and machine learning, this seemed like a good fit. But operationally, I still definitely take a lot of the viewpoints from my training as a statistician and use those in my work, even in computer science. Great. And the interest in computing in itself in order to sort of, you know, get into math, science and statistics. How did that come about? Is that something that was there from your school days? What was your first introduction to it? Oh, yeah. So, you know, I think, like a lot of people who start out on sort of a math stats path, I started out, you know, just in school doing the sorts of, you know, problem sets and things that people do in college and high school or whatever. And there wasn't a whole lot of computing. But then in college, as I realized how much I enjoyed actually applying the things I was
Starting point is 00:02:30 learning in my statistics courses to real data, it became fairly clear that computing was an important aspect of being able to actually learn things from data, except in fairly toy type of problems. And so, yeah, I think that's where that came from, just sort of marrying the interest in statistics with actually wanting to learn something from data. So the component of applied statistics is really what got me headed in this area. Got it. Yeah. And, you know, I mean, I think it's very interesting that you say applied statistics because using data in the criminal justice system is a very unique application for somebody who comes from even computing or math and stats.
Starting point is 00:03:08 How did you chance upon this field? What is it that sparked the interest for you? Yeah, you know, like a lot of my career, it's been a little bit of a winding path. And I think of this as a little bit serendipitous. So back in, I think, so 2012, 2014, I was working in this lab at Virginia Tech in the Virginia Bioinformatics Institute. This is going to seem like a little bit of an aside, but I promise I'm coming back to criminal justice here. And I was, as it turns out to be fairly relevant these days, working in this lab where we did micro simulations. So simulating things like epidemic outbreaks. And my job there was I was really focused on how you build a realistic population of agents that can pass the disease around to each other in ways that are realistic. And as I was working there, I started thinking about how much I just
Starting point is 00:03:51 how my interest expanded beyond those sorts of applications and other sorts of application areas we were working on into the sort of like social problems. So how can we use these sorts of methods to understand social problems? And so one of the things I was thinking about was applying the methods that we were using to model incarceration as a contagious disease. And so that was really my sort of entree into this area was it was really a fairly direct move in the sense that I was applying the same sorts of methods that we were using to simulate things like infectious disease, the spread of infectious disease through a population, to think about what sorts of social influence can cause close associates of people who are incarcerated to themselves become incarcerated
Starting point is 00:04:33 and looking at how racial disparities in sentence lengths could be playing a more important role than we previously thought in driving racial disparities in incarceration rates in the population. That's a whole project to explain, so I won't go into all of the details, but that was really kind of the beginning of my work in that area. So then years after that, I was working at the Human Rights Data Analysis Group. And this is a human rights organization, and we do quantitative analysis for projects that are pertinent to human rights issues. And a lot of the human rights community spend a lot of their time looking outward. So thinking about issues and things that are pertinent to human rights issues. And a lot of the human rights community spend a lot of their time looking outward. So thinking about issues and things that are happening outside of their own countries. As we were talking about it, we were thinking,
Starting point is 00:05:12 you know, we have a lot of stuff going on here at home that is certainly relevant to human rights. And so we should be trying to take this lens of human rights and statistical data analysis and all of these things and apply them to the issues that we're seeing at home. And so we started out by looking at policing and we've sort of moved our way into the criminal justice system more broadly in our work there. Wow. And the kind of data that you're looking at, I mean, this is again, maybe this is just my naivety, but do you find that the data that you have is sufficient both within the US as well as outside, for you to actually make meaningful, you know, draw meaningful insights from it? Yeah, that's actually a great question. So when I started out doing this work, I was really having to rely only on data you could find on the internet.
Starting point is 00:05:57 So luckily, a lot of cities and states are making criminal justice relevant data available on the internet for people to do their own analysis. And so it wasn't like there was a complete lack of data and we had to, you know, build things up from nothing. But it was more difficult to find the sorts of information that we needed to do really solid analysis. Now, as time has gone on, and I've gotten deeper into this field, I have been more able to come up with data sharing agreements with various institutions within the criminal justice system to have access to more private data that you're not just going to find on the internet for good reason. And so, yeah, I actually do think for
Starting point is 00:06:36 the most part, I do have the data that is necessary to do the sorts of analyses that I'd like to do. But of course, there's always things that aren't measured or things you would like to be able to understand better. But yeah, I'd say by and large, I do now based on having built up a little bit of a career in this area and having made connections in this area, so that I can get that sort of access to data, again, under fairly strong data sharing agreements and things like that. I mean, that's a great segue sort of into what I was definitely going to get at, which is when we talk about data, then, you know, of course, privacy can't be very far behind. Even in areas such as like personalization and shopping, you know, there is a sense of intrusion that you feel if the recommendations that you get are too close
Starting point is 00:07:17 to what you're studying, you know, your browsing patterns, etc. What kind of concerns have you dealt with in this specific, you know specific realm with regards to data privacy? And are there any special considerations that you need to make? Yeah, you know, so a lot of the data that you're going to find if you do start working in this area is going to be personally identifying information, right? And so you have to take any sort of precautions that you would do there with any other sort of data where you have that sort of personally identifiable information. So it's sensitive information, right? It's information about people's criminal history. In some cases, it's information about interviews that have been conducted with the people while they've been incarcerated. It's all sorts of stuff that really shouldn't entirely be public.
Starting point is 00:07:57 And so there's really good reasons for it to be private. And of course, we take all the precautions that are necessary to make sure that that remains private as we're using it. I think when we're talking about data from the criminal justice system, we also get into some really kind of like different ethical gray areas where some of the data that we have is actually a matter of public record and you can kind of find it out there in the world. It's not really super private in the sense that, you know, it is sort of on the public record. And so you can just get it via FOIAs or, or however, there are questions about how appropriate it is, I think, to publish that data with the individual's names in it, even though it's there. It sort of makes it easier for people to find information that's not really relevant to the analysis itself, but could years down the road,
Starting point is 00:09:00 I think, be embarrassing for the person or cause them more difficulty getting a job. If it's sort of out in areas that are more easily searchable and findable, say published in an academic journal or especially some sort of open access sort of paper. And so I think you get into this sort of strange gray area where there is data you can get where people are individually identified, but it is a matter of public record. So what do you do with that, right? Do you have any sort of obligation to not further put that person's name out there as sort of like tarring them as a criminal for life, right? I think these are areas that we need to think a lot about. Right. No, absolutely. And one of the other things that I wanted to touch upon, which you had mentioned previously, is while you go through and review this data and actually perform analysis on
Starting point is 00:09:44 it, you're sort of starting to draw insights from it. I have sort of a two-part question. One is, if you feel like the data is not complete enough, what has been your experience in terms of influencing organizations to either collect more data or share more data with you? And the second question I had is, as you draw these insights and uncover these, you know, potential biases that you may see, how receptive are they to your findings? How receptive are they to actually take it and make changes within their organization based on what you find? Yeah. All right. So for the first question, how successful have I been in getting people to collect more data? Not at all. So that's a really easy one to answer there. You know, there's a sort of there's a lot
Starting point is 00:10:26 of momentum around what sorts of things are collected and how they're stored. And I certainly can't go redesign all the systems that people use to store data. And so I haven't even really tried on that front. When it comes to the reception, you know, I think it's really mixed. And it really depends who you're talking about. So a lot of my work is sort of done through this critical lens that points out how unfairness in the criminal justice system ends up being inherited by models that are built on that data. And so in a lot of ways, the critiques I have of the models are critiques of the system itself, which I think can be tricky. What I've found is that the reception really depends on who you're talking about. But I've been surprised at how much people have been willing to listen and how much
Starting point is 00:11:09 I've been invited in to various conversations to sort of make these critiques heard in places that I didn't expect to have that opportunity. So yeah, it's a mixed bag, I think. Interesting. Have you at all had the opportunity to observe maybe or follow like a certain organization over a period of time, like from a time that you sort of make suggestions or you're invited in to provide your opinion to see that, you know, they've made possibly made changes and you've had an opportunity to sort of evaluate the data coming out again. Have you had that sort of a timeline to review the data? So I wouldn't say anything. I've been involved in anything that's happened for quite that long. One example of something I've been involved in was this research advisory council with the
Starting point is 00:11:53 mayor's office of criminal justice in New York City. And this group of researchers was convened to help give advice. We didn't ultimately end up being able to say you should do this or should not do that, but just give, or you can do this or you can't do that. But it was just to give advice on the redesign of a new risk assessment tool that they were rolling out and have since rolled out in New York City to assess the likelihood that someone who's been arrested will fail to appear for a court date if they are released. That was just rolled out fairly recently. So we're not really at the stage where I'm seeing data to be able to evaluate how it's performing now. And I'm not even sure I will have access to that. But I did get to see the whole design process from a fairly early stage. And it was definitely interesting. And I did feel like in several cases, they were listening to the people who are there representing points of view that weren't necessarily like traditional in terms of how risk assessments should be made. That's very, very heartening to hear. I'd like to switch gears a bit, Christiane, and go back to something that you said initially, which is that when you started
Starting point is 00:12:55 your career, you were looking at spread of infectious diseases. Given the times that we're in right now, I'd love to hear more about, I know you're working on something related to the global health situation with COVID-19. Could you tell us a little bit more about that? Yeah, I guess, you know, I've got several projects teed up on this. It might be because I'm one of those people who, when presented with anxiety-inducing situations, just work myself into the ground. So that's probably, I should probably be embarrassed by the amount of things I have going on this, given the short timeline we're talking about. But one of the projects I'm really excited about is working with some epidemiologists. So an old colleague of mine, Eric Lofgren, who's at Washington State University, and a new colleague through this project,
Starting point is 00:13:37 Nina Pfefferman, who's at the University of Tennessee, to model the spread of COVID in jails. And also this not just jails, but in jails and in the communities, because jails are a fairly porous barrier between the community and the jail itself. So people are constantly coming and going from jails. And so the top line finding that we have there is that if you can reduce transmission in jails, that has spillover effects into the community. So you can also reduce infections in the community as well. And so, you know, what sort of interventions could reduce transmissions in jails? Well, you could do things like arrest fewer people, expedite release, things like that. And so we have this model, a sort of standard epidemiological model with some additional bells and whistles
Starting point is 00:14:18 attached to it. We're actually modeling the flow of people in and out of jail and within jail to their court dates and back that looks at how the spread of COVID in jails can have an impact on the whole community. So that's something that should be coming out within the next, what's a Friday? So early next week, I would think. And I'm really excited about that project. I should also mention, and this was a huge oversight to not say this in the beginning when I mentioned the other two researchers I'm working with. I'm also working with some researchers from the ACLU, Aaron Horowitz and Brooke Watson on this project. So that's been a really fantastic collaboration that has just been all consuming for like the past three weeks or a month. And the final thing that I'm pretty excited about on this topic is I'm working on a project estimating the parameters of epidemiological models.
Starting point is 00:15:04 So like an SEIR model or an SIR model, while accounting for time lags in the data. So the idea here is that normally these types of models end up getting fit to the case count. So how many infections have we counted and over what period of time or what's the time series look like? But in this case, because in the United States and elsewhere, there's this lack of testing, right? The numbers that we see for the case counts are at least as much a reflection of how tests are being distributed as they are a reflection of the number of cases. And so we were thinking, what data do we think is fairly trustworthy that we can use? And we were thinking, well, you know, probably the number of deaths is more trustworthy.
Starting point is 00:15:42 So the number of COVID-19 attributable deaths, that's probably something we should be building a model on instead. And other research groups have come to this conclusion as well. But the problem with this is that you have this time lag from the time of infection to the time of death. People don't typically die within a day of catching the disease. So we're working on building a model that incorporates that type of time lag explicitly into the likelihood of the model so that we can come up with new ways to estimate that. The epidemiological curves, given the reality that one, we don't really trust the case count data, so we have to base the model on something else.
Starting point is 00:16:17 And two, there's this non-negligible time lag between what we get to see and what we're trying to estimate? You know, I mean, given the situation that all of us are in right now, and, you know, I think we're all sort of armchair data experts. I mean, I think we're consuming this data at a mind-blowing pace, right, both in terms of news as well as information like this, that everybody wants to know more, everybody is sort of anxious about it. So I think the work that you're doing is simply fascinating. But this idea that you mentioned of the count not being a reliable metric, is that something that you have seen? Is this primarily in the case of infectious diseases? Or are there other areas that also have this sort of challenge of not being able to get a reliable count? Yeah, so that you're actually asking a question that goes back to most of my work at the Human Rights Data Analysis Group. So I would say the bread and butter of that organization and really a lot of the work that gets done out of that is understanding
Starting point is 00:17:15 undercounts in other types of situations. So what we focus on there is casualty estimation. So in the time of some sort of violence, it can be really difficult to collect reliable data on the number of people who've been killed. And what we found in our experience there is that when local organizations, government organizations, NGOs, whatever, are trying to collect data on the number of people who've been killed, it doesn't end up being a representative or complete sample. And that's not to say they're not doing an excellent job because all of the people who are trying to collect this data in times of conflict are doing incredible work under really, really difficult situations. But the reality is, just like I was
Starting point is 00:17:54 saying with the COVID-19 case counts, that it's as much a reflection of distribution of tests as it is the number of cases. What we find in data on deaths in times of conflict is it's just as much a reflection of where resources are being allocated to do that sort of data collection as it is a reflection of how many people have been killed. And so the work I've been doing at the Human Rights Data Analysis Group for the past, I think, something like five years, has been developing methods to take lists of the names of people who've been killed from various on-the-crown organizations and apply statistical models to come up with an estimate of the number of people who weren't recorded by any of the organizations. So the sort of dark
Starting point is 00:18:34 number, the number of people who ended up being completely unrecorded. And this goes back to some ecological models that originally were developed for animal populations. So estimating the size of, say, fish in a lake, you catch fish in a lake, you tag them, you throw them back in, and then you see how much overlap there is in your second catch with your first catch. This might seem like a little bit of a harsh comparison when we're talking about estimating human deaths. And of course, we don't like to imply that there's some sort of similarity there in terms of the seriousness. But when it comes down to the statistical modeling, the methods are fairly similar in the sense that they are based on looking at the overlaps among the lists of names of people who have been killed. So if one organization collects the names of a bunch of people and another organization collects another
Starting point is 00:19:19 list of the names of people who've been killed, you can look at the overlaps among those lists. Normally, we wouldn't do it with only two lists, because that requires you to assume something like the lists are independent, which we don't believe. But as you get more and more lists, you can relax some of the assumptions you have to make. And sort of, again, to go back to, you know, the simple example, which again, this is just a simple example here. It's not actually how it's done. The intuition here is if there's a whole lot of overlap between the lists, well, that means you probably got most of the people because there's both collection efforts captured almost all the same names. If there's not a whole lot of overlap, well, then you think the universe of
Starting point is 00:19:54 individuals out there who've been killed is probably much greater than what you're observing, even on the union of the lists that you're looking at. Again, it gets more complicated, and there's all sorts of different methods you can apply to account for capture heterogeneity, to account for dependence between the lists, all sorts of things like that. That ends up being, I think, a really interesting area to do research. Well, I'm curious to ask though, I mean, in this case of the counting being inaccurate, what would you say is the biggest risk of that? Is it mostly our response to some of these situations that might not be as urgent? Like what would, in your observation, what is the biggest challenge with this and how big of a problem is it? Yeah, I mean, so I think that really
Starting point is 00:20:36 gets at why do we do this, right? Like what's the point? So in my work with the Human Rights Data Analysis Group, in many cases, we have partnered with truth commissions where the idea is that after some sort of violent conflict, in order for the peacemaking process to move forward, there really needs to be some sort of shared narrative about what happens and sort of shared understanding about who is committing atrocities and against whom. And so, you know, at the sort of the most base level, just even understanding the magnitude of the conflict, I think can help with that process of really acknowledging all the harm that had been done. And the other important, I think, reason to do this is if you can come up with different estimates disaggregated, say, by space or time or any sort or perpetrator or victim, then you can come up with a better understanding of the dynamics of the conflict. So for example, if you find that some minority group was much, much more likely to have been
Starting point is 00:21:31 killed, but not that much more likely to have been recorded, that can hint at hiding of those sorts of killings, and it can also support claims of genocide. Another aspect of this is you can look at what sort of policy interventions, what their impact was. So for example, if there's some sort of policy intervention and you find that after that intervention, the number of recorded killings goes down, it's really useful to understand, was it the case that the number of killings actually went down or was it the case that people became too afraid to report them? So the mechanisms to report safely were also decreased. So being able to disentangle the reporting from the actual level of violence, I think, is really important for that sort of retrospective policy analysis.
Starting point is 00:22:10 Absolutely. And I know that you, you know, your work in this field also extends to sort of academic as well as, you know, industry setting in the sense that I know you help organize the ACM Fairness, Accountability and Transparency Conference or the ACM FACT Conference. What is the main sort of goal of that event? And you know, what kind of attendees do you sort of attract? I mean, are there a lot of people working on the kind of problems that you're trying to solve? And is this sort of a gathering of those similar and like minds? Yeah, so this is this is something I'm really proud of. I'm really proud of seeing this community really grow over the last three something years. So the goal of this conference is mostly
Starting point is 00:22:51 encapsulated in the name. It's to study fairness, accountability, and transparency in socio-technical systems. And so this is meant to attract attendees from a whole variety of disciplines. And in fact, it does. So we have people from law, from policy, from statistics and machine learning, from sociology, from all sorts of different types of studies, philosophy, who present their work there. It's highly interdisciplinary. The other thing that I think is really neat about this conference is we have avenues for people who don't publish in the ways that academics traditionally do. So via, say, like an eight-page conference paper, there are also ways that they can have content in the conference. So we have people from advocacy organizations, their policy organizations, and they'll present
Starting point is 00:23:34 things like tutorials or these sort of interactive sessions. I participated in this really interesting session in this past one, deconstructing your memories of how you've thought about designing a sort of fair model. And so we have people from all across the spectrum, all sorts of different institutional types, all sorts of academic disciplines. And it's really a place for people to think really deeply about the impacts of technology on society. And I think it's really important to have all you know, all different sorts of perspectives represented when we're thinking about those issues. So that's the community I'm really proud of. Well, that sounds amazing. And I mean, I think that intersection of folks from, you know, various disciplines is so critical to have these sort of key conversations and actually take
Starting point is 00:24:19 forward some of these issues, especially, I mean, given that it is a conference that's run the sort of foundation of technology, I think it given that it is a conference that's run the sort of foundation of technology, I think it's amazing that you're able to attract folks from all of these various disciplines who really need to have a say in these matters. Yeah, absolutely. And one key contingent that I did leave out and shouldn't have is industry. There's also an industry presence there as well, which I think is, it's also important to be talking to those folks. Absolutely. Yeah. Speaking of folks from industry, a lot of our listeners are actually practitioners, potentially young professionals,
Starting point is 00:24:50 who also would love to hear more about you from you on just your career, right? What's your journey been like? How did you navigate your career? You come from a math and stats background into computer science, and that transition is fairly seamless. But we'd love to hear any unique stories that you may have to share from your career journey. Sure. My career has been a little bit of a winding path, in fact, more than most people's, I think. So the beginning of it starts out pretty straightforward. I went to Rice University, and I started out studying math. And about halfway through, I realized statistics, this is a thing for me. And so I went straight from there to grad school at Duke where I did a PhD in statistics. But by the time I was done with that, I was feeling fairly burned out. I was 25
Starting point is 00:25:33 and felt like a little bit of an adventure. So I took a postdoc down in Rio de Janeiro, Brazil and ended up going down there by myself at first. And after a few months, it wasn't really working out. I was ending up spending more time trying to deal with the bureaucracy than I was getting real research done. So I ended up, I just kind of left, but I ended up spending the rest of the year down in Brazil, kind of taking a break and reevaluating where I wanted my career to go and what I was going to do. I also traveled around Europe for a little bit. So I had a good bit of a breather after grad school and after the beginning of the postdoc that I ended up leaving early.
Starting point is 00:26:10 To back up a little bit chronologically, about halfway through grad school, I ended up sending a cold email to the founder of the Human Rights Data Analysis Group, Patrick Ball, because about halfway through grad school, I'd been talking to a friend about how I was really excited about all the work I was doing. And I was doing most of the things, just building models and coming up with algorithms for estimating those models, you know, sort of standard methodological statistics sort of things. I was really interested in finding a way to make those skills have some sort of social impact. So I talked to a friend who had heard Patrick give a talk. I think she was in law school and he gave a talk there.
Starting point is 00:26:43 She's like, you should really talk to this guy. So I looked him up. I was really interested in his work. He was doing a lot of this population estimation for casualties. So estimating the number of deaths and conflicts like I just talked about. And I asked her for his email and she didn't have it. So I was like, well, okay. So I think it took me about 10 Google pages in and I found his email on like the last page of somebody's CV. And I was like, you know, what the heck? I'll just cold email him and see and see what happens. And so this is about halfway through grad school. I cold emailed him and I was like, Hey, really like your work. I think I could help in these ways. What do you think? And he's like, why don't you come out for the summer? And we ended up going down to Columbia and working on
Starting point is 00:27:17 a project down there for those three months. And I sort of stayed in the orbit of HR DAG after that. So that's an important detail going forward. So anyway, after I came back from this sort of like year of just taking some time off and thinking about what I wanted to do, I ended up landing at the lab that I talked about earlier where we were doing micro simulation things. And I was designing synthetic populations to do those sorts of simulations. After about two years there, I decided that it was better for me to move back down to Durham to be with my husband because we were living separately. And just sort of for personal reasons, it was too hard to be separate for so long. So I ended up moving back down to Durham. But at the time, we had a really good friend from grad school who had just started a
Starting point is 00:28:00 startup. So our friend Wes McKinney had started this company called Datapad, which was doing basically like cloud-based big data visualization analytics sort of stuff, and they needed a data scientist. So I signed on there as employee, I think, number 10, and spent several months there before the company got acquired. So I also got some experience in, yeah, in sort of like Silicon Valley startup world. And I was traveling back and forth from North Carolina, where I was living to San Francisco to work on that. But then after that, so I told you this is a winding story, right? So sorry. It's fascinating.
Starting point is 00:28:37 But then after that, gosh, this must have been like 2015 or so. And I'm sure I could get the exact dates, but I don't think that's important for the story. Yeah, so it was around 2015. After that all ended, I ended up staying and doing tech consulting just because I'd made a bunch of connections in that area while I was doing this sort of startup thing. So I did some data science consulting for some small startups, like one was called Treasure Data. I also did some data science consulting for larger companies. One of those is called eBay, where we were looking at building an anomaly detection system. And I also spent about half my time back with HRDag because I felt like that was
Starting point is 00:29:13 sort of where my heart had been since grad school. I'd really felt like that was something I was really interested in pursuing again. And so I was sort of splitting my time doing all of that. And then I think sometime around 2016, I ended up, it was a lot to be doing a bunch of different types of consulting. It's a lot to keep track of when you're doing things that are just so different. And so I ended up joining full time at HRDAG. I think it was sometime around 2016 and stayed there until just now, about a month ago when I moved over to the University of Pennsylvania.
Starting point is 00:29:41 So a whole lot of different types of things from, you know, the sort of standard trajectory, then into a research lab type of environment, then to a startup, then to sort of being an independent data science consultant, and then sort of back to the nonprofit world and now back to academia. So it's kind of this full circle type of journey. Yeah, that's great, Christiane, because you're like sort of the poster child for somebody that we, you know, we think about when we say transition from academia to industry and back is actually very challenging, or just even working across those lines seems to be very difficult in many cases. But your journey has
Starting point is 00:30:19 been proof of the fact that it can be done. You know, I think it's especially inspiring, because you started out this entire journey with a cold email. And sometimes we're so afraid to make those decisions and say, should I even reach out? And I think your example is a great way for us to think about this and say, you know, what is the worst that can happen? You won't hear back. Yeah, that's about it. That's great. So, Christiane, we're sort of running out of time, but I'd love to hear more about like, you know, now that we're all cooped up in our respective homes and sheltered in place in many locations, what do you do in your free time if you have any?
Starting point is 00:30:52 You know, I don't have a whole lot of free time. So what I do these days, since we are all cooped up, is, well, I would be doing this anyway. Let's be honest. I'd be doing this anyway. I hang out with my daughter. I have a 16-month-old daughter who takes up a whole lot of time and energy right now, but I wouldn't have it any other way. She is just a delight and a joy. And I don't need to go on on that because I think probably most parents feel that way. And I could go on for hours about how wonderful she is. And that would just be beside the point and probably kind of annoying. Some of my other hobbies though, which I don't get a lot of time to do these days, maybe in a couple of years, I really like rock climbing, mostly indoors, some outdoors, especially when I was at Virginia Tech, I used to really like to go
Starting point is 00:31:33 rock climbing outdoors in the River Gorge. And the other thing that's kind of weird, I guess, is I really like sewing costumes. So I used to, and I plan to revive this again, go to the Renaissance Fair at least once a year and pretty much always in a new costume and so that was sort of my yearly project was sewing something brand new for me my friends my husband any number of people who wanted to go wow yeah that's that is a unique hobby have you been sewing any masks because I know a lot of us have. I plan to, I have not yet, but I definitely plan to. Got it. Well, it's been amazing, Christiane, talking to you. For our final bite, I'd love to hear what is it that you're most excited about in your field of computing over the next five or so
Starting point is 00:32:17 years? Yeah, I feel like I haven't really had, as you can tell from my like, what was my career like answer, I haven't really had much of a plan. I've sort of always just followed wherever I felt there was something interesting to do at the time. And so I think that's probably what's going to keep happening for the next five years. I'll probably just continue following whatever's, whatever piques my interest as long as I can, you know, have a job that lets me do that. So yeah, I don't know. Right now COVID-19 is really interesting. Fairness is really interesting. That is fairness in machine learning. There's all sorts of things that are really interesting and we'll see. We'll see what happens. Great. Thank you so much for talking to us. It's been an absolute pleasure
Starting point is 00:32:58 to host you on our show. Thank you for taking the time to speak with ACM ByteCast. Thanks so much for having me. ACM ByteCast is a production of the Association for Computing Machinery's Practitioners Board. To learn more about ACM and its activities, visit acm.org. For more information about this and other episodes, please visit our website at learning.acm.org slash ByteCast. That's learning.acm.org slash B-Y-T-E-C-A-S-T.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.