Microsoft Research Podcast - 059 - Building contextually intelligent assistants with Dr. Paul Bennett
Episode Date: January 16, 2019The entertainment industry has long offered us a vision of the perfect personal assistant: one that not only meets our stated needs but anticipates needs we didn’t even know we had. But these uber-a...ssistants, from the preternaturally prescient Radar O’Reilly in the TV show M.A.S.H. to Tony Stark’s digital know-and-do-it-all Jarvis in Iron Man, have always lived in the realm of fiction or science fiction. That could all change, if Dr. Paul Bennett, Principal Researcher and Research Manager of the Information and Data Sciences group at Microsoft Research, has anything to say about it. He and his team are working to make machines “calendar and email aware,” moving intelligent assistance into the realm of science and onto your workstation. Today, Dr. Bennett brings us up to speed on the science of contextually intelligent assistants, explains how what we think our machines can do actually shapes what we expect them to do, and shares how current research in machine learning and data science is helping machines reason on our behalf in the quest to help us find the right information effortlessly.
Transcript
Discussion (0)
If you think about having your own executive assistant, a great executive assistant is
prepared to answer all of your questions.
And that's where a lot of the assistants live right now.
They're responsive to you.
They know the answers.
They come back.
An even better assistant starts anticipating your needs.
And that's where we go with things like the Calendar Aware Proactive Email Recommendation.
You're listening to the Microsoft Research Podcast,
a show that brings you closer to the cutting edge of technology research
and the scientists behind it.
I'm your host, Gretchen Huizenga.
The entertainment industry has long offered us a vision of the perfect personal assistant,
one that not only meets our stated needs, but anticipates needs we didn't even know we had. But these Uber assistants from the
preternaturally prescient Radar O'Reilly in the TV show MASH, to Tony Stark's digital know-and-do-it-all
Jarvis in Iron Man, have always lived in the realm of fiction or science fiction. That could all
change if Dr. Paul Bennett has anything to say about it. As principal researcher and the research manager of the Information and Data Sciences group at
Microsoft Research, he and his team are working to make machines calendar and email aware,
moving intelligent assistants into the realm of science and onto your workstation.
Today, Dr. Bennett brings us up to speed on the science of contextually intelligent assistants,
explains how what we think our machines can do actually shapes what we expect them to do,
and shares how current research in machine learning and data science
is helping machines reason on our behalf in the quest to help us find the right information effortlessly.
That and much more on this episode of the Microsoft Research Podcast.
Paul Bennett, welcome to the podcast.
Thank you.
You're a principal researcher and you're the research manager for information and data sciences at Microsoft Research.
So you wear at least two hats, maybe more.
Yes, that's right.
Tell us more generally than specifically right now
what you do for a living.
What big problems are you trying to solve?
What gets you up in the morning?
Within the Information and Data Sciences group,
we think about two things.
First off, information science is about how people work with
and use information in every aspect of their lives. Data science,
we think about how do we analyze the data and behavioral traces we have from people
interacting with data? How can we predict their actions? How can we understand what they're trying
to do? What we try to do in the group is actually bring these two together and understand how people
are working with information, how their behavioral traces say what they're doing and what they'd like
to do, and what software may not even support them in doing right now that they'd like to do
even better. And so, you know, really one of the things that I think about at a very high level of
what the group does, it's about taking science fiction and making it reality. So things that
you've seen in movies that, you know, 15, 20 years ago you said are not possible, first off it's
about making those real, and then it's about envisioning the future and laying the groundwork for how we get to the science fiction of the future.
Before we get into the meat of the interview, I want to set the stage by talking a little bit about recommendation engines writ large.
It's sort of a huge part of our life now and a part of the science. And we humans seem seriously invested in developing tools that
will predict our next movie, our next purchase, our next turn, our next relationship. So I'd love
to get your take. What's our core problem here in making choices and what are researchers doing
about it? Right. Well, let me unpack that a little because I think you addressed two different things
in there. One is predicting what actions you might take. Another is helping people make decisions. And there are actually two interesting parts that come into artificial intelligence. In terms of recommendation, the variety of recommendation engines that have been developed, you're right, have been targeted at things like advertising or selling you products. Even going back to my days as an undergraduate, my first project was actually
on content-based recommendation engines.
Turns out, looking back at it at the time,
I was actually writing a web crawler
to scrape Amazon and recommend books.
This was back in 97 and 98,
when Amazon was developing their own recommendation engine
at the same time.
At the point of time then, that seemed very novel
because people hadn't thought of it applied to any task.
Now, people have applied it to a variety of things to sell you products. But the next question is, you know, for intelligence,
even when we interact with people, you're thinking about what's going to happen next. How will they
react to what you're going to say? You're trying to predict the future. And in some sense, that's
what recommendation engines are about. Can we predict the future and actually guess what your
next action will be? But more importantly, and this comes to the decision-making question, often the
challenge is you don't know all the possibilities and when people are trying
to interact with a set of items or a set of possibilities, they don't know all the
different capabilities that are out there.
So an example of this, you know, just from search engine point of view, if you went
and searched for a flight on a search engine and said, you want to fly from
Seattle to Austin, and it you want to fly from Seattle to
Austin and it came back to you with a single result at your price preferences at the single
best time and showed you this one result, the question is, are you happy? And if you stop and
think for a while, what most people come to the conclusion of is, no, I'm not happy. And the reason
is you don't know what else is possible. And that's where decision-making comes in.
Decision-making is not only about finding what matches your interests,
it's about knowing what the other possibilities are
and that this is a good decision for you to go forward with.
And so when we think about presenting people with information,
we have to give them enough supporting information
that they understand where they're going from there.
I made a note of a statement I read in a blog on this topic that email and calendar recommendations
are underexplored in AI research. So what was the aha moment or inflection point where
you decided to do an AI research project to recommend emails instead of shoes to me? So I think there were a couple of small moments
leading up to things. One was exactly looking to this question of how people make decisions and
what supporting information they need. But then we were doing work specifically trying to understand
where the gaps are in intelligent assistance of when you interact with a person, what types of
common sense does the person have that the system is lacking? And one of the key conclusions that we came to was that
it was lacking a lot of information about context. And there are a variety of things
that go into context, but some of the key variables are time, people, and what it's about.
If I know when something is going to happen and where you think it's going to happen,
if I know the people that are involved and I know what it's about, those things always exist
with email and with calendars or nearly always.
The data's there.
Sometimes you have very useless descriptions of calendar items like sync or talk.
How many sync emails do you get?
It's frighteningly high actually.
But you know, they were just creating it because it was in the path of communications
of what they're sharing. But the information that it gives the system then to reason on their behalf
and actually help them proactively, that just offered a huge opportunity.
Reason on their behalf. That's such a big statement. Can machines do that?
Absolutely. In very small, limited ways right now. And of course, what we do in research is
push that boundary even further. But, you know, when I think about reasoning on someone's behalf,
what I think about is, first off, you want to be able to predict what they're going to need
and help them out. So if you think about having, you know, your own executive assistant,
a great executive assistant is prepared to answer all of your questions. And that's where a lot of
the assistants live right now. They're responsive to you. They know the answers. They come back. An even better
assistant starts anticipating your needs. And that's where we go with things like the calendar
aware proactive email recommendation, which is I have this meeting. I didn't actually remember
what information was relevant. You were able to determine it and you were able to predict I was
going to search for this. And instead of wasting my time during the meeting, boom, it's just there.
But then when you really talk about reasoning on someone's behalf, you start anticipating
what are they going to wish they would have thought of even after this event where they
came to realize they missed an opportunity, the things that you could have actually put in there.
So for example, someone sends me a design document for some other
project. Doesn't seem relevant at the time. I kind of ignore it, go on with things. I have some other
planning meeting for a new project, completely different. Turns out that that other design
document that I ignored actually has relevant information. Worse yet, it's going to create a
blocking situation six months down the line. I have all the right information at my hands, but
of course I didn't have the time. And at the time that I received the first document,
there was no reason to think that it was relevant. And when you talk about reasoning on someone's
behalf, it's about saying, look, that information is there. You already have access to it. Can we
bring that to your attention and actually say, hey, maybe there's a little part here that if
you dig into can prevent huge problems in the future.
So I'm intrigued by the technical underpinnings and research methods that might help people like me find, as you phrase it, the right information effortlessly.
Take us into the technical weeds for a minute.
Are you using existing methodologies in new areas or are there some fundamental explorations in ML methods?
Yeah, so we make use of both.
And we also even go beyond that within the group.
We try to first take an HCI perspective and say,
if we were to try to identify what it is that people need,
sometimes how they interact with the system
is heavily influenced by its current capabilities.
And so how do you even get after what they would use
if those capabilities were there or what those needs are?
So for example,
coming back to this question I mentioned about time and not understanding time, we said, what is
it that people want a system to understand in terms of time? And so we said, well, what would
it be like if you interacted with another person? So we can go out and do surveys like that. The
problem is there's all sorts of biases to that. So we say, what other types of data are available?
So here we had email as a huge opportunity where you're communicating with another person.
So we can actually look at how you communicate time to another person and what types of language do you use?
As opposed to when you go and set a reminder on, you know, whatever intelligent assistant you use right now, you constrain your language because you think the system is only capable of certain types of things.
And so here's the interesting thing. When you go to an intelligent assistant,
the first thing that they want to do
when you tell them something is say,
what time should I remind you?
And it's just this problem, which is,
I don't know what time.
And so the interesting part of this is that
we can actually capture even this kind of soft information
so that we can come back to you proactively.
And it turns out as you get near that point in time,
you can often resolve that further. More interestingly, if you actually do
interviews, which we followed up from looking at how people communicated an email in this way,
it's not just about that item. Turns out when you made that note, you probably were doing something
else. And your first goal was actually take the note quickly. Let me continue to focus on this
item. Right. And so that's one of the things that some of the researchers in our group look at is how do we allow people to focus and be in the
moment of what they're doing and realize that what they may currently be trying to take a quick note
on or do something else is not actually the key thing. So that's a set of looking from a very
user-facing perspective, understanding where the gaps are. Now let's talk predictively. What are
we trying to do in terms of prediction?
So there we have information traces that we can work with.
So for example, you interact with certain types of meetings
in different ways.
You send a meeting invite out to somebody.
You may reference different links in it,
have different attachments.
Those are all indications of what's relevant
to that meeting.
And so that gives you an opportunity to automatically say, here's some examples of some things that are relevant to that meeting. And so that gives you an opportunity to automatically say,
here's some examples of some things
that are relevant to this meeting.
Then we can look across all of your meetings,
emails and documents and say,
can we predict on a regular basis then what is relevant?
And so there we would come
with typical machine learning techniques
where you can start with very simple techniques
like simple linear logistic regression.
You can move on to more complicated techniques,
nonlinear methods like gradient boosted decision trees,
neural networks.
It turns out that in this space,
the key challenges though are not so much the architecture
of the model that you're looking at,
but how you deal with the privacy challenges,
which is very, very again,
different than things like advertising and selling you books.
In that case, everybody is buying the same books.
In this case, only you have access to your information, which means the traces
that we learn from are only relevant to you. And how we think about how we generalize across users
in ways versus within a particular user are key challenges. Yeah, so I'm going to come back to the
what keeps you up at night and what are your challenges with privacy a little bit later in
our interview,
but I'm hearing two kinds of research going on.
One is the user research, which is super important
to find out what the needs are and what the pain points are.
And then you're over here on algorithms and machine learning
and the deep technical weeds of how you take
what you've learned and make a machine be able to do it.
Yes, that's right.
So part of it is HCI, or human computer interaction focused. And even within that
field, there are huge challenges in understanding the biases that you deal with when you try to
elicit from people what their preferences are, how they'd interact with things. But there's also
actually an interesting part that's in between two that connects them. So if you look at a current
system, again, how you expressed a reminder to an assistant to give you a certain time, you constrained yourself based on what they expected.
And we see the same thing in every single system.
So, for example, in search engines, people don't type questions that look like a question quite often for two reasons.
First off, typing it takes a while.
But there's a second reason that's more interesting.
They don't actually issue those queries because they don't expect the system to answer them.
Because they know from experience that those don't lead to success. And so as we add the capabilities that the systems are capable of answering those kinds of questions, there's
challenges getting people to use them and to actually, you know, teach them the behavior.
It's all about what if, what if something different would have happened? And so we see the same thing
in any given system.
What if we would have given you a different set of documents and said,
these were relevant to this meeting, would you have interacted with one of these or not?
And so there's a line of research within the group looking at counterfactual or
causal reasoning that tries to say, okay, this is the data we have, but what we want
to reason about is the system we're going to deploy, which is going to behave differently
and how you're going to react to that.
So how do we take those past behavioral traces and deal with the different biases that
are introduced in terms of how you reacted because of that system, as opposed to how you will react
to the new system? So how are you doing that? So there are an interesting set of techniques,
and this is, again, you know, new novel techniques being developed as well. Some of the simplest methods look at
actually doing small amounts of randomization in terms of minor things. So if you think about
the order of documents you're presented, if you add just the slight amount of randomization
so that you don't put them in order of your predicted relevance, but a slightly different
order, you're now actually able to reason over a broader set of orders that you didn't put in front of people, right?
And even small tweaks like that allow you to make changes.
Now, one of the interesting challenges is where we made progress deal with very limited situations like that.
Rankings, an item from a small set of items.
But as we go more and more to rich interfaces, there are all sorts of questions of how does the presentation affect things. So even simple things like email, email when you do a search may look
like a ranking of items, but then it turns out there are conversations. These conversations can
be collapsed or not. They can have different types of interface markup. All of these affect
your likelihood of interacting with those items based on more attentional factors,
but that influences the data that we learn from. And so being able to separate out how your attention influences your
interaction versus what we really want to estimate, the relevance and pertinence of the information
to you in order to update the models is one of those key challenges. You have to be willing to
kind of understand that there may be a suite of different things that are most appropriate to the
situation, as well as being willing to push on the boundaries and develop new techniques
when appropriate and finding out where you can actually deal with these different kinds of
challenges by adding a slight amount of noise to the process or even showing in you know in other
cases we've done research where we looked at and said how much can we introduce randomization
without creating an experience that you'd hate, right? Because you could imagine that you actually went
into a restaurant to look at a meal. I could put in front of you five choices that you really liked
where I could add a random choice just to learn about what you might like that you didn't know
that you liked, right? You might be willing to tolerate that one out of five. Now you walk in
and you have five random choices in front of you and a horrible dinner, right? I may have learned a lot about you,
but I can't also, you know, I can't run that risk. And so being able to understand the limits of how
we can do that online exploration, but also sometimes we can't always do online exploration.
We have just what we've observed in the past. And so there we try to identify what are natural experiments,
where there's sort of small amounts of variation
due to random differences in the system.
A lot of times we have distributed systems in the cloud
where things are returned slightly differently
because of latency.
And those offer natural experiments
that allow you to say,
can I learn from what just happened to be random chance
and better predict the future? And a lot of it comes down to when you learn a model
that is influenced by these presentation or selection biases that happen, you deploy this,
and then what you actually find out is it doesn't work. And so it's a false start,
and you may repeat and repeat and repeat. And so when you learn from the logs to more effectively
estimate what will happen once you deploy it, it's also about rolling out innovation to our customers much more quickly.
Efficiency is a huge thing in computer science research, making us more efficient.
And when I talked to you before, you said it's not specific enough.
You know, efficiency is a step on the path of where we want to go. It's certainly not the end goal.
Ultimately, intelligence is about changing outcomes for people, giving them capabilities
they wouldn't have otherwise had, or allowing them to essentially be augmented by the system.
And so when I say that, what I mean is that when we make intelligent predictions for people,
we assume that there has to be some reason for them to select what we've put in front of them. The first initial case is usually in terms of efficiency. So, for example, you know, if I'm retrieving this information relevant for your meeting, this was going to be information you're going to search for anyway. However, in order to issue that search, you would have to say, you know, from Paul about some particular topic. And then it turns out we've exchanged
email about a number of things and you have to find out a way to cut that down.
And which email from Paul is the one I remember?
And so I've made it efficient for you to identify that, right? But when we think about changing
outcomes, so this is the case, you know, where I mentioned the design document, that design
document, you didn't even know was relevant. You wouldn't have searched for this. If I actually
can put that in front of you and allow you to ask relevant. You wouldn't have searched for this. If I actually can put that in front of you
and allow you to ask questions
that you wouldn't have otherwise asked,
more generally, when we think about augmenting capabilities,
we think about things like, what do our expert users do?
How do they interact with the system?
How do we take a novice and turn them into an expert, right?
So you imagine just starting out your job,
you're a fresh college graduate,
you're interacting with Visual Studio, maybe a few other tools you've been searching around.
You know, you have your favorite websites for development that you go to, but there's a wealth of other resources you don't know.
When we can connect you with those and suddenly accelerate your learning curve as a developer because of the patterns of other people, that really changes your life. It's not just about efficiency.
It's about changing what information you have at your hands based on how you're interacting.
Let's talk now about this project that you're working on called CAPERS.
It's Calendar Aware Proactive Email Recommender System.
So tell us about CAPERS.
What's new, cool, different about it?
How does it work?
Who's using it?
How's it been received?
Yeah.
So in terms of the system itself, it's just this notion that when we look at how you go
throughout life, there's a lot of different meetings that you may have. The typical professional has
three or four meetings a day. And one of the complaints, of course, is they have too many
meetings. But even if we can't just get rid of their meetings, the question is, how can we
actually make their information life better with respect to their meetings? And so part of the
things that we see is that people have trouble finding information to prepare for a meeting. They
have trouble during a meeting, finding information on demand that they didn't anticipate needing.
They have trouble after a meeting, referencing back to things relevant to the meeting.
And when we look at meetings, we also see again that there's context, right? There's the context
of the people that were involved. The topics usually isn't the meeting name, the time. And so all of this gives us an
idea of when you'll need this, what phase are you in? Is it before the meeting and you're likely
preparing? Is it during the meeting and you're really trying to respond in the second, which
means that you might be looking for different types of information. Again, it's information
you were less likely to anticipate. Is it after? In which case you're looking for follow-ups
that were sent out after the time of the meeting
more likely than before.
And so this really offers us the capability
to understand the phase,
the types of information you want,
and to predict those things seamlessly.
In terms of where it is right now,
the Microsoft Search and Assistance Group
is looking at productizing this.
This particular project is being led up by Abhishek Arun, but he's also working tightly with the user experience folks
in Microsoft Office and FAST, which is a group that we have in Norway. So right now, this system
is being used internally by a small set of people. We roll things out in phases and you get to start
seeing, you know, how much are your assumptions right as people interact with things? How often
do you need to update the model because what people need evolves?
Some of these questions I mentioned in terms of presentation and selection biases actually do come up again, because if you have only certain types of positive information, like what was relevant to this meeting, that's biased by what people's search capabilities were at the time.
And so there's a tendency to go just towards what they currently were able to find.
And so being able to expand that set, all of those things come up as you push it further and further.
But yes, the system is usable right now.
And hopefully in the future, we'll have it actually in customers' hands as well.
And so you've used it.
I've used it, yes.
What's your experience?
I mean, is it radar-esque for you now?
It's radar-esque.
I think there's certainly times where I'm just wowed by it, where I didn't expect something and it was there.
You know, actually, the more amusing thing, and it's a matter of using the future too,
I had this colleague of mine who's leading up Abhishek Arun come give a talk and asked him for the slides that he presented afterwards. And he said, why are you asking me? Just open up your meeting.
And so, you know, you also have a, how do you change your own habits as well?
But even serendipitous cases where that was something where he had shared access with me,
but not sent me the file, but it was there in terms of being relevant to the context. And so
those capabilities where you can wow people, those are really fun chances. What will be fascinating for
me if I use it is you can sort through the number of emails that say quick question that I sent or
follow up. And that's the entire subject, right? And I can't remember what I said or what I was
following up on. So how you present content in terms of giving
people a quick summary, that is a key challenge. There's, you know, a part of this that I haven't
talked about that Ryan White and Eric Horvitz and I have worked on with Cortana and again,
folks in Microsoft Search and Assistance, the Cortana suggested reminders where it looked at
your email and said things that you would take action on in the future. So, you know, this is
the kind of case where my wife sends me an email and says i got the theater tickets for saturday night
can you get the babysitter and i respond and say yes i'll get the babysitter somebody walks into
my office i forget about it saturday night comes and didn't get the babysitter exactly and so if
you pop up a small reminder and say to people hey it, it looks like you said you do this. Do you need to block out some time or is it already done?
Just that simple item there can actually make them take action.
But coming back to your question of what kind of information do you need to give them,
you have to give them enough context where they know what it's about.
Speaking of that, CAPERS is just one example of where contextual assistance can come into play.
Where else and with whom else are you digging around
in this research bin where you say good ideas
and deep research understandings
turn into product-ready features?
Yes.
So I mentioned kind of the Cortana suggested reminders
as another example,
and that's tied into the same space
of how we understand how communications relates
to the work you're trying to get done.
And that offers key opportunities where we partner with both other groups in research,
like the Language Information Technologies Group and the Knowledge Technologies and Experiences
Group, as well as our product partners to make those things happen.
More generally, even going back to old assistants in the research space like Lumiere that Eric Horvitz was driving
back in, I think, the late 90s. One of the lessons we learned is that in order to provide people
assistance, you have to be within their current workflow. If your primary application that you
have open and you're using right now is Word, then we need to think about how to integrate with Word.
And that's part of thinking about contextual assistance in general. What does it mean to provide assistance in that moment? So other kinds of cases we're looking at,
also partnering with groups in office as well to look at applications such as PowerPoint.
How do you think about and reuse content across different types of documents? So you might have
different slides that you need at one case. How do you refine them and reuse those later?
These kinds of opportunities come up in application after application. I mentioned understanding
how people search within the office suite. How do we go from search directly to actions?
How do we recommend those features in general? Those come up across the whole different suite
of products.
What about other verticals like education or, you know, these other big chunks, healthcare,
et cetera? Yes. Great question. For the most part, we haven't done much within the group with Like education or, you know, these other big chunks, healthcare, etc.
Great question.
For the most part, we haven't done much within the group with healthcare a little bit.
Education comes a little bit more closely to things that we've been looking at.
So we have partners in Windows and Office together with external partners at Harvey Mudd and the University of Michigan looking at what will the reading experiences of the
future look like?
And in particular, how does information about people's attention that we can take from things like gaze
tracking or cursor movement, which is a proxy for gaze, help us understand where people's attention
are, where they're reading, what they're learning or what they may not be learning? How can we
change that reading experience to enhance their learning? And I think that offers an interesting
capability in general. I mean, even right now, if you look at teachers, again, teachers don't really know
what their students are or are not paying attention to. And so this is the case of really
thinking about how can we augment those capabilities.
That, Paul, was a perfect segue into my next topic here, which is, as I think about all the
data you'll need to collect about me in order to anticipate my needs, including needs I don't even know I have, and all the ways you're going to
collect that data, my eye movements, my cursor movements, my documents, my emails, et cetera,
I get a little nervous about what I'm trading off and who's looking at it in exchange for this
amazing convenience of effortlessly finding my information. So is there anything about what you
do that keeps you up at night? I think that there are a host of different
privacy challenges and the kinds of methods that we can develop within these spaces versus what we
know is the best possible just in terms of prediction. So privacy is always foremost in
our minds. We don't look at customers' data. Algorithmically, we may run over the data,
but how we even generalize from that data is limited based on our privacy guarantees.
So, you know, think of it like in terms of the algorithm running where I learn from your documents what's important to you. I have to be careful that if I'm going to learn about
particular words or phrases that are important to you that I don't transmit to anyone else indirectly through the model that those words or phrases are important to you, right? And so even
simple cases like imagine that I knew acquire LinkedIn was a more common phrase in Satya's
email box. That would be terrible, right? It doesn't matter even if it was available within
Microsoft. It would be terrible. And so, you know, we think about how do we construct guarantees on privacy between our different customers
that prevents anything from going in between, right?
Even indirectly.
And so, yes, absolutely, the first thing we do is say,
you know, we will limit ourselves
in terms of the predictive capabilities
to make sure that we guarantee that first.
Which is an interesting promise and perspective. So how can you educate users,
customers, et cetera, that this is actually true, what you're saying? It is safe. You're
not going to look in my underpants. You know, I think we just have to,
first off, keep communicating that, but then also just be true to our word,
not just in what we do, but in who just be true to our word, not just in what
we do, but in who we partner with. You know, we work with a lot of external partners. We ask
our partners, how are you working with your data? Who agreed to this? Was this a broad enough
agreement? Why would we work with any partner that can't meet our own ethics and guidelines
on what we think is acceptable? And so I think, you know, being very clear across the board that this is not just a subject of policy that we work within because we've been told,
but deep beliefs we hold is foremost. Tell us a bit about Paul Bennett.
What's your background?
How did you end up working in information and data sciences at Microsoft Research?
I started quite a while ago, originally looking first at just philosophical problems of how we reason
about things. And I started... Was this when you were a child? How far back are we going?
Not that long ago. But going back to undergrad, and I was actually looking about how do people
reason about court cases? So I not only did philosophy, it was more from the theorem proving
side, the logical side of things. And so in undergrad, I did both philosophy and a computer science degree where I looked
more at the computer science side again of doing this.
And you know, the computer science part of the project I mentioned was to actually build
this recommender system as well on the computer science part of things.
But more generally, as I got deeper and deeper into AI, it was seeing how AI could really
transform situations and moments. And for me, probably the biggest moment
I had with artificial intelligence was I was working on a machine translation project to
translate from English to French. And I just finished coding up this particular algorithm
we'd been talking about. I'd worked out what I knew were the software systems bugs, and it ran.
I typed in English, and I got French out, but I looked at it
and I said, I don't know if it's right. This was interesting. And I was actually sitting in an
international conference in Finland at the time. And because of the international conference,
there's a number of people around. And I raised my hand and said, does anybody speak French here?
Somebody said, yes. And I walked over to them and said, did they say yes or wait?
I walked into them and said, does this French sentence say what the English sentence says? And they said yes. And then I typed another one. Does this say the same thing? They said yes. And I did
another sentence. Yes. And then I realized I'd written a program that was capable of doing
something that I couldn't do. And that is simply magical. When I can just from an algorithmic specification and data create something that, you know, I wasn't capable of, at least in the moment, without lots of study, that kind of capability to be able to make it available for all the moments in our life, that's just a huge opportunity.
So how did you end up here? After undergrad at the University of Texas, I went to graduate school at Carnegie Mellon. During my time in grad school at Carnegie Mellon, I actually came out as an intern. And so, you know, if you're an undergrad or a graduate student listening, internships are a hugely important part of your career development.
Who did you work with? with Eric Horvitz and Susan Dumais. At the time I was looking at how do we actually do ensembles
of machine learning methods. That gave me an opportunity to start engaging with them. And then,
you know, when I went back to Carnegie Mellon, I continued to engage with them throughout the
time of my dissertation. And then after I was finished with grad school, I came out and started
as a researcher. Carnegie Mellon seems like an incubator for MSR in a lot of ways. A lot of
people have come. It's also,
I don't think it's completely coincidence in that going back several decades, Carnegie Mellon said,
computing cannot just be about computer scientists. It has to be about how computing changes the rest of the world. And so they started actually bringing in other disciplines at the time
and saying, what happens if we put people together from different disciplines? How would it apply to this portion of the world? I like to ask all my
guests when we come to the end of our conversation to offer some parting thoughts to our listeners.
And it can take a number of forms, whether it's, you know, advice or challenges or inspiration or
next steps. How might some of the brightest minds think about joining the efforts in this area, Paul? I think there are a number of different ways that people can contribute.
First off, you know, in terms of us, we're open to working with people. We bring visiting
researchers through, we start different partnerships. And certainly if you have someone
that you'd like to reach out to here, feel free to reach out to them and more than likely they'll
get back to you and kind of talk about opportunities. But I think that more generally, part of the computing mindset is to say,
where could computing be that it's not? You know, what are the opportunities? Something that you
look at and you say, this is not an opportunity for computing. Usually that's not the case. It's
a question of how would you see it differently? How would you turn it into an opportunity?
Where is the current friction? You know, I even think about
two simple things like during undergrad, I worked for the government for a while and
we had these forms, everything had to be on an official form. And you know, at the time
they didn't even have this where they could print from a software system to this form
within the office. And one of my great innovations at the time was I created templates where you could photocopy
onto these official forms most of the things
that you needed for each person in the office.
And that again was just an example
of the computing mindset, right?
It didn't look like a computing opportunity.
I couldn't create a word processor for everything there
because they wouldn't be the official forms.
But it was a question of how do you actually recognize
the generality and reapply it?
And that can happen in any field,
regardless of what you work in, regardless of whether
you're an AI or not, you know, dreaming about the possibilities and then finding a few people
who might actually have what you don't have in those other areas to bring it together.
And you say, hey, I can bring this from this field.
What do you have from yours?
Let's talk about where we can actually go together.
Paul Bennett, saving time and changing outcomes.
Thanks for coming on the podcast today.
It's been great.
Thank you, Gretchen.
To learn more about Dr. Paul Bennett
and how researchers are building
contextually intelligent assistants
to deliver contextually intelligent assistants,
visit microsoft.com slash research.