Microsoft Research Podcast - 118 - Accessible systems for sign language computation with Dr. Danielle Bragg

Episode Date: June 17, 2020

Many computer science researchers set their sights on building general AI technologies that could impact hundreds of millions – or even billions – of people. But Dr. Danielle Bragg, a senior resea...rcher at MSR’s New England lab, has a slightly smaller and more specific population in mind: the some seventy million people worldwide who use sign languages as their primary means of communication. Today, Dr. Bragg gives us an insightful overview of the field and talks about the unique challenges and opportunities of building systems that expand access to information in line with the needs and desires of the deaf and signing community. https://www.microsoft.com/research

Transcript
Discussion (0)
Starting point is 00:00:00 As machine learning becomes more powerful, having data to train those models on becomes increasingly valuable. And in working with minority populations, we're often working in data-scarce environments because the population is small and there might be other barriers to collecting data from those groups in order to build these powerful tools that actually can really benefit these minority communities. You're listening to the Microsoft Research Podcast, a show that brings you closer to the cutting edge of technology research and the scientists behind it.
Starting point is 00:00:35 I'm your host, Gretchen Huizenga. Many computer science researchers set their sights on building general AI technologies that could impact hundreds of millions or even billions of people. But Dr. Danielle Bragg, a senior researcher at MSR's New England lab, has a slightly smaller and more specific population in mind. The some 70 million people worldwide who use sign languages as their primary means of communication. Today, Dr. Bragg gives us an insightful overview of the field and talks about the unique challenges and opportunities of building systems that expand access to information in line with
Starting point is 00:01:15 the needs and desires of the deaf and signing community. That and much more on this episode of the Microsoft Research Podcast. Danielle Bragg, welcome to the podcast. Thank you. It's great to talk to you. I like to start by situating both my guests and their labs. And let's start with your lab. You're a senior researcher at MSR New England, and you're hanging out in Cambridge, Massachusetts. Tell us about the work going on in the lab in general and why it's important. What gaps or research gaps do the folks in Cambridge fill?
Starting point is 00:02:01 Yeah, we're located in Cambridge. The lab is a very interdisciplinary place to be. We have a lot of people from different fields, not just computer scientists, but we also have economists, people who work in theory and social science, which makes it a really interesting place to work. We work on problems that are both technically challenging and have societal impact. So the collection of skills that we have really fits that mission. We have machine learning experts, economists, theory folks, as well as social scientists all working side by side, which makes it a really
Starting point is 00:02:39 rich place to work. Yeah. Well, let's situate you now. Many of my guests locate themselves at an intersection. Where do you live, Danielle? What are your particular research interests and passions and what gets you up in the morning? Yeah. So my work lies within human computer interaction or HCI, but it also fits under accessibility and applied machine learning. So kind of the intersection of those three, I would say. What is particularly interesting to you right now as you are looking at the sort of broad picture of the research you're doing? My work primarily focuses on building systems that expand access to information, in particular for people with disabilities and in the last few years, in particular for sign language users.
Starting point is 00:03:31 So I'm a computer scientist. And as we discussed, I fall under the general umbrella of HCI, but also touch on other fields. And I would say I'm personally very motivated by problem solving and by working on problems that I feel have some positive impact in the world. Well, before I talk to you earlier, I have to admit I had some big gaps in my understanding of what sign language actually is and how it works and perhaps some of our listeners are the same. So give us a bit of a primer on this very singular language. How and why does ASL stands for American Sign Language, for those listening who haven't heard the acronym before. And ASL is a natural language.
Starting point is 00:04:35 It has its own grammar and vocabulary, just like English or any other spoken language. There are actually many different sign languages used around the world, which some people may not realize. And American Sign Language is the primary language of the deaf community in the United States specifically, as well as a few other areas around the world. There are a number of linguistic features that make up sign languages, just like there are linguistic features that make up spoken languages. For ASL, there is handshape, location of the hand on the body, and movement. Those are the three primary types of features, but there are a whole host of other features that are also important. For example, non-manuals, which include facial expressions and other types of body gestures. There's fingerspelling, classifiers, depictions where you're kind of acting out certain content.
Starting point is 00:05:32 It's a really beautiful language, and there's a really rich culture centered around it, just like there's rich culture centered around other languages around the world. So interestingly, I wasn't even going to ask this, but as you bring it up, I'm thinking to myself, are there idioms within the language? Is there slang within the language? Are there things that are outside the normal sort of structural grammar of the language as it evolves with people and generations? Yeah, yeah, there definitely are.
Starting point is 00:06:04 There are different dialects used by different subpopulations. There are also just really rich genres of literature. There's deaf poetry. There's certain types of stories that people like to tell. And because the language is visual, there's a lot of richness there that you don't really get with spoken languages. But I should also give a disclaimer. I'm not deaf myself, and I'm still learning a lot about this space. I've taken some ASL classes and learned about deaf culture and the deaf community. But, you know, I don't have a lifetime of experience. So I'm always learning as well.
Starting point is 00:06:40 Just as a point of interest, is there Chinese sign language? Is there Spanish sign language, French sign language, or is it that granular? Yes, there's Greek sign language, British sign language, French sign language. There are many different sign languages across the world. However, American sign language is actually more closely related to French sign language than it is to British sign language. So the relationships between the sign languages don't always mirror exactly the relationships between spoken languages. And that's because, you know, there's a different history and evolution and the groups of people who are using those languages mixed in slightly different ways. But they are basically geographically situated because
Starting point is 00:07:26 people who physically live near one another talk to one another more. Right, right, right. I got my mouth open just like, I didn't know that. I didn't know that either. We're off to a great start. Well, before we get into your technical work, I think it's really important to understand who you're doing it for and why you're doing it. And we've sort of alluded to that already, but when we talked before, you mentioned two groups, deaf signers and people who have hearing loss but don't sign. So you're addressing two sort of populations there. More interestingly to me, though, is the fact that you frame this whole thing in terms of culture. So I'd like you to talk about your goal of improving technical accessibility for the two main groups and how that plays out. But maybe you could help us understand the cultural aspects first. So the deaf community has a really rich culture and ASL is a very important part of that culture. And in this conversation, we're focusing on ASL because we're here in the U.S., and that's where most of my work is focused, but a lot of this applies to other sign languages as well.
Starting point is 00:08:31 And within the deaf community, ASL has a really sacred place, I would say. It's a really beautiful language, and it's kind of the binding glue for the community in many ways. And a lot of my work focuses on helping to preserve ASL and supporting people who want to use ASL. So a lot of my work is about supporting sign language use and supporting people using it in their interactions with technology. Being deaf is a point of cultural pride for many people. So many people who are deaf don't view themselves as disabled. They view being deaf as a cultural identity. If you're deaf, you can still do anything that anyone else does. You can go to the store, you can drive a car, you can go to work,
Starting point is 00:09:25 but the communication piece is where the barriers come into play. And communication is central to culture, right? So people who share a language develop cultural connections with one another in a different way. Well, you've put the big picture of building accessible information systems in a data-driven frame. So talk about this approach writ large and how it's informing the specific projects and papers you're working on since data is central to the technological approaches that many of you are working on right now? Yeah, data is central to a lot of technologies that are being developed. As machine learning becomes more powerful, having data to train those models on becomes increasingly valuable. And in working with minority populations, we're often working in data-scarce environments because the population is small and there might be other barriers to collecting data from those groups in order to build these powerful tools that actually can really benefit these minority communities. And so in my work, I try to build data-driven solutions.
Starting point is 00:10:35 And in doing that, I often will try to actually collect data in a system that is also providing benefit to the community. So we don't have to go to the community and say, oh, give us your data, we'll pay you, or provide some other kind of compensation. If we can actually build systems that provide benefit to the community while they're contributing, that can be a much more organic solution to this type of problem. Okay. If you're approaching this from a data-driven perspective and data is scarce, what's the biggest problem you face in your research right now? Well, I would say one of the biggest challenges is dealing with this data scarcity and figuring out how to collect data in these environments actually presents a host of really rich research problems to work on. You can be really creative in designing systems that
Starting point is 00:11:33 incentivize people to participate and provide benefit while also collecting data to then train other models and provide other types of services. Well, let's go upstream for a second and talk about what kinds of models you want to provide that you would need this data for. So what kinds of sort of top level applications or solutions are you aiming for? Yeah, so within the sign language space, the dream in some sense would be to provide end-to-end translation between, for example, English and American Sign Language. And that translation needs to be bidirectional, right? So it's not enough to just recognize signs and translate that into English. We also need to let the deaf person know what, you know, people speaking English around them are saying.
Starting point is 00:12:23 So we need to translate from English to American Sign Language as well. And recently, there have been some advances in deep learning and convolutional neural nets in particular that seem promising in this space. But it's important to note that any technical solution would be dictated by the needs of the deaf community and would not be a replacement for human interpreters. Let's talk about what you call sign language computation, which is sort of an umbrella term for all the research going on here. Give us an overview of the current state of the art for sign language computation. And then, and this is going to be a multi-part question, so I will keep bringing us back,
Starting point is 00:13:16 making sure we cover everything. Talk about the biggest challenges you face in five areas that you identify as data sets, which we've sort of already talked about, recognition and computer vision, modeling and NLP, avatars and graphics, and then UI UX design. That's a lot to unpack. If we get lost, I'll bring us back. But let's start with the state of the art of sign language computation. Sure. So that breakdown into those five groups is really helpful for thinking about this space. So those five areas are really needed for developing end-to-end bi-directional translation. So first, we'll talk about datasets. Existing sign language datasets
Starting point is 00:14:02 are primarily in video format, and there are a number of different ways that people have tried to collect these videos. You can curate videos from professional interpreters. You can try to scrape different online resources. But these are all limited in some way. In particular, the diversity of the signers in the videos, how many deaf fluent signers you get, as opposed to students or professional interpreters is also limited often. And just the sheer size of the data set is also very limited. So to put that last problem in context, for speech corpuses, we typically have data sets between 5 million words and 1 billion words large. And for sign language data sets, the largest data sets we have are less than 100,000 signs total. So that's a very large difference in how much data we have. And if you think about the history of speech recognition, how long it took them to get to where they are today, and how much difference having all that data has made,
Starting point is 00:15:10 that might put into context for you how hard this is. Okay. So if we're talking about data sets being limited and you're looking for robust machine learning models to help get to a robust sign language computation application. How do the other things play in? You mentioned recognition and computer vision. Let's talk about that for a second. Yeah, so in the space of recognition and computer vision, for sign language recognition,
Starting point is 00:15:38 it's a pretty young field dating back to the 80s when people used hardwired circuits and rule-based approaches, for example, fitted to gloves that had little sensors in them. Those types of systems are limited in how well they work. In addition to the technical constraints, gloves also have other problems. So if you're using gloves for recognition, you're missing a lot of important grammar information that is on the face, for example. And you're asking someone to carry around gloves and put them on all the time. And they also don't provide this bidirectional translation that's really needed to have a conversation, right? If you're wearing gloves and signing, maybe some microphone can
Starting point is 00:16:25 speak out what you're saying, but then if someone talks back to you, you have no idea what they're saying. So it's a very incomplete solution. But for technical reasons, people started out with those types of approaches. More recently, advances in neural networks, for example, CNNs and hybrid models that pull together information from different types of models have been promising, but we're still operating in this data-limited environment, so we don't actually know how well those models might perform given enough data. All right, so the recognition in computer vision's state-of-the-art isn't very good state-of-the- art is what you're saying. Yeah, basically. And so the challenge for researchers there is what can we do instead or how could we augment or advance what we've done in these areas with new tools, new approaches?
Starting point is 00:17:19 I mean, yeah, people are playing around with different types of models. People have also tried to be clever with pulling together multiple data sets, for example, or tuning parameters in certain ways. But ultimately, my intuition is that we really need more data. Once we have more data, we can figure out how to finesse the models, but we don't even know how far the models can take us right now because we don't have the data to fully try them out. All right. Well, I want to get back to how you're going to go about getting data because we had a really interesting conversation about that a couple days ago. But let's continue to unpack these five areas. The next one we talked about was modeling and NLP, natural language processing.
Starting point is 00:18:00 How does that play into this? Yeah, so modeling and NLP is very important for figuring out how to translate and how to do other interesting computations with sign language. These types of approaches have traditionally been designed for spoken and written languages, which introduces certain difficulties. For example, there are certain assumptions with written and spoken languages, in particular that one sound happens at a time. But in signed languages, one movement doesn't always happen at a time. You can have multiple things going on at the same time. And some of these models don't allow for those types of complexities that a sign language might have. Another complexity is that the use of space can be contextual in sign languages.
Starting point is 00:19:01 So sometimes if you point to the right of you, you might be referring to yourself at home. And another point, while you're talking to someone, you could reestablish that area to mean yourself at the coffee shop. And so we need to have contextual models that can recognize these types of nuances. And the models built for speech don't account for these types of complexities. So we may need new types of models. Another big problem in this space is a lack of annotation. So even if we have videos of people signing, we often don't have written annotations of what is actually being signed. And a lot of the NLP techniques, for example, really rely on annotations that computers can process in order to work. Okay, these are huge challenges. Well,
Starting point is 00:19:53 let's talk about avatars and graphics as another challenge in this field. Yeah, so avatars and graphics are needed to render content in a sign language. So we've talked about this bi-directional translation that would be great to facilitate. And in moving from English to ASL, for example, you need some kind of rendering of the signed content and avatars and computer graphics provide a nice way to do that. The process of creating an avatar is actually really complex, and right at the moment, a human is needed to intervene at basically every step of the way. So we have a lot of work to do in this space as well.
Starting point is 00:20:39 But typically the process starts with some kind of annotated script that gets translated into a motion plan for the avatar. A number of parameters then need to be tuned, for example, speed within individual signed units or across signed units. And then finally we need some animation software to actually render the avatar. I should also mention that avatars have had mixed reception among the deaf community, especially if they are not very realistic looking. They can be kind of disturbing
Starting point is 00:21:13 to look at. So there are lots of challenges in this space. Are they sophisticated enough to even get to the uncanny valley or are they just lame? I mean, it probably depends on the avatar. I suppose. Well, either way, it sounds expensive and cumbersome to have this be an approach that's viable. Yeah, it is difficult. I mean, there are some companies and research groups that have tried to make avatars and they typically spend a lot of money collecting very high quality examples of signs that they can later string together in the avatar. But even with that, you need a human to come in and manage and clean up whatever is generated. Well, let's talk about UI UX design and that interface between deaf signers and computation? What are the challenges there?
Starting point is 00:22:07 So I think UI UX design is another really rich space for exploration and development, in particular because sign language is a different modality from written and spoken languages. But again, a big challenge here is designing interfaces that will be useful despite our lack of data and despite the limitations that our current technologies have. So figuring out ways to provide a human in the loop solution or provide results that are good enough that can then learn from users as they're using the system or other types of creative ways to support users becomes a really rich space for design and exploration. Right. So there's a lot of opportunity for research in this area and probably a lot of
Starting point is 00:23:01 room for other researchers to join the efforts. Yeah, definitely. I think it's also one of the most interdisciplinary spaces that I've come across. You need people who are experts in deaf studies and linguistics and HCI and machine learning. You need all of these areas to come together to make something that's really going to be useful for the community. Tell me a little bit more about your ideas and approaches for actually gathering data. You've alluded to some of the difficulties in the existing data sets. So how might you broaden your data collection? Yeah, so that's a great question. I can give an example of one system that I've been working on that both provides benefit to the community and collects useful data at the same
Starting point is 00:23:54 time. So one project I've been working on, it was started when I was a PhD student at University of Washington with my former advisor, Richard Ladner there, is to build an ASL dictionary. So if you come across a sign that you don't know, and you want to look it up, that can be really challenging. Existing search engines and search interfaces are typically designed around English, but it's really hard to describe a sign in English. And we also just don't have videos indexed that way, right? Like what would your query look like? Right hand moves right, left hand moves up, you know, two fingers extended.
Starting point is 00:24:33 We just don't support those types of queries. And also searching by gesture recognition also doesn't work very well because we don't really have the capabilities working accurately yet. So we designed a feature-based dictionary where you can select a set of features that describe the sign that you're trying to look up, for example, different types of hand shapes or movements. And then we match that against a database of past queries that we have for science in the database and sort the results based on
Starting point is 00:25:07 similarity to past queries in order to give you a good result. And in this way, while you're using the dictionary to look up a sign, you're actually providing data that can be used to improve the models and improve results for people in the future. So these types of systems where users are providing data that will actually improve the system going forward can be a really nice way to jumpstart this problem of data scarcity. Right. And you talked earlier about existing data sets, which involve maybe some videos that have been taken from somebody giving
Starting point is 00:25:45 a speech and having a deaf signer in front or beside. And are those all public domain? Are you able to use those kinds of things that exist and just pull them in? Or is there a problem there as well? Yeah, that's a great question too. So some data sets are public domain, some are not. So collecting sign language data is very expensive, not only in terms of dollars spent, but also in terms of time and resources. And so groups who collect data sets may be a data set, but it could also be companies who are trying to build a translation software and they're trying to outdo their competitors. So there are a lot of data sets that are not publicly available. We don't actually know exactly how big those data sets are because they're not public, but it seems like they're pretty small based on the quality of existing translation and recognition systems. All right, well, let's move on to a recent paper that you published. In fact, it was in 2019, and it won the Best Paper Award at
Starting point is 00:27:07 Assets. And you address many of the things we've talked about, but the paper also addresses the problem of silos and how to bring separate portions of the sign language processing pipeline together. So talk about the questions you asked in this paper and the resulting answers and calls to actions. It was called Sign Language Recognition, Generation, and Translation, an Interdisciplinary Perspective. Yeah, so we were trying to answer three main research questions. First is what is the current state of the art of sign language technology and processing? Second, what are the biggest challenges facing the field? And then third, what calls to action are there for people working in this area? And as you mentioned, this is a very interdisciplinary workshop in February of 2019.
Starting point is 00:28:07 We invited a variety of academics working in a variety of fields. We also had internal attendees who are employees at Microsoft. And in particular, we made sure to invite members of the deaf community because their perspective is key. And they led a variety of panels and portions of the day. And as a group, we discussed the five main areas that we have already talked about and kind of summarized, you know, what is the state of the art? What are the challenges and where do we go from here? So that paper was presenting our results. All right, so drill in a little bit on the siloed approach and what some of those problems are as you work towards a robust application in this arena. So I think we touched on this a little bit earlier when I was talking
Starting point is 00:28:59 about some of the challenges in using NLP techniques for sign language computation. A lot of the NLP techniques are developed with spoken languages in mind, and so they don't really handle all of the complexities of sign languages. So that's an example of a situation where we really need linguists or deaf culture experts combining with natural language processing experts in order to create models that actually will apply to sign languages, right? If you only have NLP people who are hearing, who use English building these models, you're going to have very English-centric models as a result that don't work well for sign languages. And, you know, the people probably don't realize that they don't apply. Right. And which gets to the question of why don't work well for sign languages. And, you know, the people probably don't realize that they don't apply.
Starting point is 00:29:46 Right. And which gets to the question of why don't you just use English? Well, because it's a different language, right? Right. Exactly. American Sign Language is a completely different language from English. It's not signed English. So if you know English, that doesn't mean that you can understand ASL easily. And if you know ASL, that does not mean that you can necessarily read English easily either. So that's a point that I think not a lot of people recognize that English in a lot of cases is a person's
Starting point is 00:30:19 second language. They can grow up signing in the home and then learn English as a second language at school. And as anyone listening who has learned a second language knows, it's not as comfortable most of the time. Let's talk about your tool belt for a minute. You've situated yourself at the intersection of AI and HCI, leaning more towards HCI, and much of your research is building systems, but you still face some of the big challenges with enough data and good enough data, as we've talked about. Talk about the research methodologies and technical tools you're using and how you're working to tackle the challenges that you face. Yeah. So as you mentioned, I do do a lot of systems building. I do a lot of website building, full stack
Starting point is 00:31:06 engineering. I mean, there's a whole set of skills that go into that. As far as data collection goes, I've used a lot of crowdsourcing, whether that be on an existing platform like Mechanical Jerk or building a new platform to collect data in other ways. We also incorporate a lot of applied machine learning techniques in the dictionary, for example, that I was explaining. Our backend is powered by latent semantic analysis, which basically does a big dimension reduction on the feature space to figure out which dimensions are actually meaningful in completing the search. I also do a lot of user studies, interacting with users in a variety of ways, and engage in a number of design practices that incorporate key stakeholders. So in particular, securing research partners who are
Starting point is 00:32:02 deaf, but also engaging in participatory design and other ways to engage with the community. I like a combination of qualitative and quantitative work. I guess that's kind of a catchphrase these days. Right, right, right. Let's project a bit and speculate how the work you're doing for the deaf community might have a positive, if unintended, impact on the broader population. Some people call this the curb cut effect, where something that was supposed to help somebody ended up helping everybody, or populations they didn't expect to help. You know, the curb cut was for wheelchairs, turned out to be great for strollers and cyclists and people rolling
Starting point is 00:32:42 suitcases and everything else. So, do you have any thoughts on other application arenas that face similar challenges to sign language computation? One thing that comes to mind is dance annotation. I have a background in that, and it's full-body expression as well. It's funny that you mention dance because there are a lot of parallels there. In particular, sign languages actually don't have a widely accepted written form, and that causes a lot of the barriers to using our text-based interfaces in a sign language. And a lot of the same problems apply to dancers, right? If you're a dancer or a choreographer and you want to write down the dance that you are coming up with or the dance that you're dancing, that can be really hard. And as a result, there's a woman who came up with a system called dance writing.
Starting point is 00:33:36 And that system has actually been adapted to create a written form for sign languages called sign writing. So there definitely are a lot of similarities between you know dance and signing and I would say more generally any gesture based human interaction has a good amount of overlap with sign language research. So gesture recognition in particular has a lot of similarities to sign recognition. I would say that gesture recognition is actually a simpler problem in many ways because there's no grammatical structures to understand, and the context doesn't change the meaning of a gesture the way it does to a sign in many cases. So gestures might be for a person on a runway who's bringing the plane in or something or what you would do with cyclists and what those gestures mean.
Starting point is 00:34:29 And they're pretty simple and straightforward. Yeah, exactly. Or you could think about interacting with a computer through a simple set of gestures or an Xbox. I know there have also been research projects to try to support people learning how to play a particular sport or do yoga more effectively by detecting gestures that the person is making and helping to correct them. Or how you learn a musical instrument, for example. The types of gestures that you make make a big difference. So I think there's a lot of overlap with other areas where human movement or gesture is involved. Danielle, we've talked about what gets you up in the morning, but now I have to ask what keeps you up at night. And you could call this the what could possibly go wrong question.
Starting point is 00:35:14 Do you have any concerns about the work you're doing? And if so, how are you addressing them up front rather than post-deployment? deployment? In all the projects that I do related to sign language, I really do my best to include perspectives from people who are deaf and give deaf people a platform to be heard and to participate and expand their careers. But that is something that I consciously think about and sometimes worry about. I personally am still learning about Deaf culture and the Deaf experience. I don't have a lifetime of experience in this space. I've taken some ASL classes, but I'm not fluent. I'm also not Deaf, and I don't have the Deaf-lived experience, so it's particularly important to include those perspectives in the work that I'm doing. And I have a number of really wonderful collaborators
Starting point is 00:36:05 at Gallaudet University, at Boston University, at Ross Chester Institute of Technology, and a number of other places. So that's what I'm doing to try to help with this, you know, with this concern. Right. What about data collection and privacy? That's a great question as well. I do worry about that. In particular for sign language data, it's a very personal form of data because the person's face is in it, their body is in it, the background, you know, if it's their home or their workplace or wherever their signing is in it. So there are a lot of privacy concerns involved in this space. I've done some preliminary work exploring how we might be able to impose certain types of filters on videos of people
Starting point is 00:36:53 signing, for example, blurring out their face or replacing their face with an avatar face. Of course, if the movement is still there, if you know the person very well, you might still be able to recognize them just from the types of movements that they're making. But I think there are things that we can do to improve privacy at least. And it seems like a very interesting, rich space to work in. Well, it's story time. What got young Danielle Bragg interested in computer science and what path did she take to follow her dreams and end up working at Microsoft Research New England? So in my undergrad, I studied applied math. Growing up, math was always my favorite subject and I still enjoy mathematical-oriented
Starting point is 00:37:37 work. Towards the end of my undergrad, I didn't know exactly what I wanted to do, but I wanted to do something practical. So I decided to go to grad school for computer science. It seemed like a practical decision. But in grad school, I was really searching for projects that had some human impact and that hopefully were making a positive impact in the world. And that's where I really got interested in accessibility. So I met my former PhD advisor, Richard Ladner at University of Washington, and he introduced me to the field of accessibility. He got me taking ASL classes and working on some problems in the space that I'm still working on today. So did you just fall into a job at Microsoft Research or were you an intern? Is that the typical pathway to the job or how did that happen? I did intern at Microsoft. I've interned at Microsoft three times actually. Once in the Bing search group and then two times as a research
Starting point is 00:38:39 intern with Adam Kalai in the New England lab. And then I did a postdoc at the New England lab for two years. And now I am a full-time researcher in the lab. So I can't go anywhere else. I'm forever a New England researcher. Awesome. What's something we don't know or might not suspect about you? Maybe a character trait, a life event, a hobby side quest. And how has it impacted your life and career? So I've spent a lot of time training in classical music performance, actually. I played the bassoon pretty seriously through college and considered being a professional musician at that point.
Starting point is 00:39:24 I studied with someone in the Boston Symphony and went to summer music festivals, which is a thing that pre-professional musicians do in the summers. And I still have a lot of friends and acquaintances in orchestras and playing chamber music. And I would say music really added a lot of richness to my life in addition to my love of music. I think my professional training actually had a lot of richness to my life. In addition to my love of music, I think my
Starting point is 00:39:45 professional training actually had a lot in common with my research training. So training to be a professional musician takes a lot of practice and dedication. And it's more of an apprentice model. So you usually study closely with one teacher at a time. And they really teach you, you know, how to play, how to make reeds, if your instrument requires reed making. And actually being trained to do research is quite similar in a lot of ways, right? You have your PhD advisor who you work closely with, and you learn from doing research alongside them. So I didn't plan it originally, but I think that, you know, being trained as a classical musician probably actually helped me a lot with training to do research. I love that. You know, there's such a huge connection between music and math, by the way, that so many researchers I've talked to have had that musical interest as well, but not in the classical bassoon playing category. So you are unique in that.
Starting point is 00:40:46 Yeah, bassoon is a different one. I grew up, my mom had a record of Peter and the Wolf, and all the different animals were represented by the different instruments, and I remember the bassoon, but I can't remember the animal it was associated with. I'll look it up after we're done. I think it's the grandfather, but I could be wrong. Well, as we close, and I'm sad to close, as we close, this has been so much fun. I've taken to asking what the world might look like if you're wildly successful. And some people frame this in terms of solving problems that would impact millions or billions of people. But I think sometimes the goal is less grandiose
Starting point is 00:41:25 and the impact might be more meaningful to a smaller population. So at the end of your career, what do you hope to have accomplished in your field? And how would you like life to be different because of your research? Well, it might sound a little cheesy or cliche, but I really hope to leave the world
Starting point is 00:41:44 a little bit better than it but I really hope to leave the world a little bit better than it was when I started out. And in my career, I hope I'll have helped people get access to information that they may not have had access to beforehand. I think education is so key to so many things. You know, not only degrees that you get from schools, but your personal development, different types of skill development, or just general understanding of the world. And I think if you don't have access to information, that's really, really a problem, right? At least if you have access to the information, you can decide whether you want to consume it, you can decide what you want to do with it. And you have the possibility of learning or advancing yourself. But if you don't even have access, then, you know, what can you do?
Starting point is 00:42:30 So a lot of my work is focused on increasing access to people who use languages that are not often served or supported or have difficulty accessing information in different ways. Danielle Bragg, this has been really great. I have learned so much from you and I'm so inspired by the work you're doing. Thank you so much for coming on the podcast today. Yeah, thank you. To learn more about Dr. Danielle Bragg and the latest in accessibility research efforts, visit microsoft.com slash research. And for the record, it was the grandfather.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.