Behind The Tech with Kevin Scott - Behind the Tech: 2020 Year in Review
Episode Date: December 9, 2020How can we use technology to better the lives of everyone on the planet? That’s a theme running through our favorite conversations of the year – from artificial intelligence to synthetic biology. ...Listen in to learn a few cool things about our extraordinary guests! View the transcript. Kevin Scott Behind the Tech with Kevin Scott f4l3zKfbAt5enScc7FUG
Transcript
Discussion (0)
I think it's really important for all of us to sort of share our stories because, you know, there are interesting lessons in there.
But maybe most importantly, understanding each other a little bit better shows us how I do honestly believe that we are vastly, vastly more similar than we are different and that sometimes these stereotypes that we have of each other are
just obstacles that stand in the way of each of us achieving our full potential.
Hi, everyone. Welcome to Behind the Tech. I'm your host, Kevin Scott,
Chief Technology Officer for Microsoft. In this podcast, we're going to to Behind the Tech. I'm your host, Kevin Scott, Chief Technology Officer for Microsoft.
In this podcast, we're going to get behind the tech.
We'll talk with some of the people who've made our modern tech world possible and understand what motivated them to create what they did.
So join me to maybe learn a little bit about the history of computing and get a few behind-the-scenes insights into what's happening today.
Stick around. Hello, and welcome to a special episode of Behind the Tech. I'm Christina Warren,
Senior Cloud Advocate at Microsoft. And I'm Kevin Scott.
And today we're doing our year-in-review episode, and this means that we're going to revisit a few fascinating conversations
with our guests from 2020 with topics ranging from artificial intelligence to synthetic biology.
I mean, 2020, wow. It's been, I think I can say, an unprecedented year, full of challenge,
but also opportunity. Kevin, if you had to come up
with a tagline for 2020, what would it be? You know, every time somebody says that
2020 is an unprecedented year, I always go back to the scenes in The Princess Bride with the Sicilian, where he keeps saying inconceivable.
And at some point, Mandy Patinkin's character says,
I don't think that means what you think it means.
Yeah, I don't know what my tagline, I mean, like my children have declared 2020 the worst year ever.
But, you know, I think even though it has certainly been perhaps the most challenging
year that I have experienced in my adult life because of the pandemic, because of this
reckoning that we've been having in public with a bunch of social issues that
we should have contended with, addressed, and resolved a long, long while ago.
I do think that it has also been a year of unbelievable progress on several different scientific fronts and where we're getting an
opportunity unlike any we're likely to ever have again in our lifetimes to think very very
carefully about our lives how we show up in society and for us in particular in the technology industry, how it is that we
map what it is we're doing on to like what are an increasingly set of obvious needs that society,
the citizens of the world, the planets are going to need in the future.
Yeah, no, I think you're exactly right. And I'm glad that you kind of
turned that whole worst year ever thing into something that could be a little bit more hopeful
because I think you make really great points that a lot of people have really stepped up and we've
really had to start to address things that maybe we haven't before. And that is reassuring and maybe
a little bit hopeful. Yeah. And the, you know, other thing, too, that I will say is that we, in a time of crisis like this,
we all have an opportunity to pay very, very close attention to what's going on,
because the things that we get to observe now hopefully won't happen again for many, many good reasons.
Like, we should never wish another
pandemic on ourself. But because of what's going on and the urgency with which we are
trying to address these problems, some things are just moving with extraordinary speed. And,
you know, that's been one of the themes of this season as we've
talked to more people doing work in the biosciences. No, you're exactly right. And so,
throughout this remarkable year of 2020, we have had quite the lineup, as you mentioned.
We spoke with synthetic biologist Drew Indy and neuroscientist Tom Daniel, with Daphne Kohler
about digital biology, and Oren Etzioni about
artificial intelligence. And we also had an insightful chat with Microsoft Research's Eric
Horvitz. And of course, Greg Shaw interviewed you about your amazing book, Reprogramming the
American Dream, which explores how we might ensure that AI better serve us all. We also met science
fiction writer Charlie Strauss,
who talked about what a real frustration it is
to come up with an original idea in times like these
when truth really is stranger than fiction.
Yeah, so, so true.
And I'm really grateful for the many extraordinary guests
we've had on this show over the past year.
I sometimes have to pinch
myself because I get to talk to such great people and have such awesome conversations. If I had to
pick a common theme from all of these conversations, I'd say it centers around the very basic question
of how can we use technology to better the lives of everyone on the
planet? And you can see this across the board. So, you know, the folks who are working in the
biological sciences are using technology in these incredible ways, like where unprecedented might
actually be the appropriate adjective to describe some
of the ways that they are accelerating progress in this field by taking this intersection
of the biological sciences and automated experimentation and high-precision instruments and artificial
intelligence. It's just this incredible mixture of things that really is helping us solve problems
that we haven't really been able to solve in quite these ways before.
And then, you know, you look at Oren Ezzioni, who is at the Allen AI Institute,
which is just doing really incredible work trying to build more powerful
artificial intelligence systems in service of solving some of the big problems that we
have on the planet and will have for years to come.
And then, you know, obviously, I always love chatting with Eric Horvitz, our chief scientist. Eric has one of the most brilliant and
interesting set of experiences of anyone I know.
I don't know too many people who have a PhD in computer science
and are medical doctors
and have spent their entire career leading a research institution like
Microsoft Research.
And so getting his perspective on things and looking at his career path was extraordinary and wonderful.
And then Charlie is one thoughtful about thinking about the human condition in the context of
all of these sort of interesting technological phenomena and phenomena shaped by technology.
Yeah, I agree. I mean, I think that it's been fantastic to hear from
so many smart people to think about the work that they're doing and to think about the potential
that that work is going to have on all of us. Yeah, indeed. So with that, let's get started.
First, we'll hear from digital biology and machine learning pioneer Daphne Kohler. I think one of the very, very thin silver linings around this very dire situation that we find ourselves in is that there is, I hope, a growing appreciation among the general public for what science is able to do for us today and how much of that ability rests on decades of basic
science work by many, many people, much of which is publicly funded work at academic institutions,
that without that level of progress that we've made, the concept of, say, creating a vaccine in 12 months would have been completely ludicrous a few years ago.
Yeah, so Daphne says it well.
She is actually a computer scientist by training. As a Stanford faculty member, she began working in machine learning with traditional data sets and quickly realized that she wanted to pursue data sets that were more richly structured and more aspirational.
So that's the reason she got into biology and medicine.
And Daphne's work is fascinating in part because it relies on the collaboration of scientific fields that historically have spoken
different languages. And Daphne describes her work at In-Citro, the company she founded,
which applies machine learning to the research and development of pharmaceuticals.
So the premise for what we're doing really emerges from what I said a moment ago,
which is that this last decade has been transformative
in parallel on two fields that very rarely talk to each other.
We've already talked about the advancement
on the machine learning side
and the ability to build incredibly high accuracy,
predictive models in a slew of different problem domains
if you have enough quality data.
On the other side, the biologists and bioengineers have developed a set of tools over the last decade or so
that each of which have been transformative in their own rights,
but together they create, I think, a perfect storm of large data creation, enabling large data creation
on the biology side, which when you feed it into the machine learning piece can all of a sudden
give rise to unique insights. And so some of those tools are actually pretty special and
incredible, honestly. So one of those is what we call induced pluripotent
stem cells, which is we being the community, not we at In-Citro, which is the ability to take
skin cells or blood cells from any one of us, and then by some almost magic, revert them to the state that they're in when you're an embryo,
in which they can turn into any lineage of your body. So you can take a skin cell from us,
revert it to stem cell status, and then make a Daphne neuron. And that's amazing because that
Daphne neuron carries my genetics. And if there are diseases that manifest in a neuronal tissue, you will be able to potentially examine, assay those cells and say, oh, wait, this is what makes a healthy neuron different from one that carries a larger genetic burden of disease.
And so that's one tool that has arisen.
A different one that is also remarkable is the whole CRISPR revolution
and the ability to modify the genetics of those cells
so that you could actually create fake disease,
not fake disease because it's real disease,
but introduce it into a cell to see what
a really high penetrant mutation looks like in a cell. And then commensurate with that,
there's been the ability to measure cells in many, many, many different ways where you can
collect hundreds of thousands of measurements from each of those cells. So you can really get a broad perspective on what those cells look like rather than coming in with, I know I need to measure this one thing.
And you can do this all at an incredible scale.
So on the one side, you have all this capability for data production.
And on the other side, you have all this capability for data interpretation. And I think those two threads are converging
into a field that I'm calling digital biology, where we suddenly have the ability to measure
biology quantitatively at an unprecedented scale, interpret what we see, and then take that back and write biology, whether it's using CRISPR
or some other intervention to make the biological system do something other than what it would
normally have done. So that to me is a field that's emerging and will have repercussions that
span from, you know, environmental science, biofuel, bacteria, or algae that do all sorts of funky
things like suck carbon dioxide out of the environment, better crops, but also importantly
for what we do, better human health. And so I think we're part of this wave that's starting to emerge.
And what we do is take this convergence and point it in the direction of making better drugs that can potentially actually be disease modifying rather than, as in many existing
drugs, just often just make people feel better but don't really change the course of their
disease.
That was Stanford computer science professor and CEO Daphne Kohler talking about her work at her company in Cetro.
Yeah, and we should also note that the recent awarding of this year's Nobel Prize for Chemistry went to two women responsible for the development of CRISPR, which is, for folks unfamiliar, the first precision technology allowing human beings to alter the genome of humans or other organisms.
Yeah, I was so glad to see that award.
Next, we'll hear from another Stanford PhD,
Drew Indy. Drew is a member of the Stanford University bioengineering faculty, and his research teams have pioneered amplifying genetic logic, rewritable DNA data storage,
reliably reusable standard biological parts, and genome refactoring. Here's Drew.
So we're biology, right? Everybody listening to this is biologic. We all have biology. So,
so, you know, like biology is kind of important as a gross understatement. So how do we explain the fact that we tend to take biology for granted?
And I think it's because, well, we just get biology. And so there's a way of thinking about
the living world, which is the living world exists before us, and we are a part of it,
and we inherit it. And we can't do anything about it. It's just, it is what it is. And before the
mid-19th century, not only is it is what it is, but it is what it is.
It doesn't change.
This is the pre-evolutionary view.
Now, post-Darwin and colleagues, we have another cultural perspective on biology.
It exists before us, and it is what it is, but it changes over time through this evolutionary
process.
And we all know well that that's controversial still, culturally, for some, right?
Do I have the pre-evolutionary view of biology? the, but, but from my point as an engineer, it doesn't really matter. Everybody
in either of those tribes is just the living world. We just take it for granted. It is what
it is. And it's not that we don't care about it, but we don't really think about it as this
substrate, as this type of material. And then a generation ago, starting 1970-ish, we get first
generation genetic engineering, and now
we're getting second-generation genetic engineering. And suddenly, we get to inscribe human intention
into living matter very crudely at first, but we're getting better at it. And so this is something of
our time. This third reality that we can express and inscribe human intention in living matter is
really a third cultural perspective. And it forces us to confront, ultimately, what do we want to say? And what do
we wish of our partnership with the living world, to the extent we can partner with it, to the
extent that we can take responsibility for our writing, so to speak. Now, what are people good
at? People are very good at caring about people. And so, of course, health and medicine are a big deal.
But it doesn't stop there.
And when I take a look at what's going on, like just to get some numbers out in the conversation,
how's biology powered?
Well, right now it's mostly powered by photosynthesis.
Well, how much?
And the answer is 90 terawatts, plus or minus.
70 terawatts of photosynthesis on land and 20 terawatts in the oceans. What's 90 terawatts? Well, civilization's running on what,
20 terawatts these days, plus or minus? So, okay, that's interesting. The energy powering
the natural living world is four and a half times the energy consumed by the human civilization.
Huh. Now, you asked me, like, what's the big deal?
How about civilization scale flourishing? Like, because what's biology doing with those joules,
that energy coming in? It's organizing atoms, right? So biology is operating at this intersection
of joules, the energy, the atoms, the material, and bits, by the way, right? The DNA code,
which is abstractable to information. And so we've, and bits, by the way, right? The DNA code, which is abstractable
to information. And so we've got this stuff, this living matter, it's atomically precise
manufacturing on a planetary scale operating at almost 5x civilization. What should we be looking
at? Lots of individual things, of course, vaccines here and there, a big deal. But the big prize I
would submit for consideration is civilization scale flourishing, where we can provision for 10 billion people rounding up
without trashing the place. And that's never been true before, because we've never understood
biology and the way we're approaching it, both as a science and engineering discipline.
And if I go away, right, and if all of bioengineering goes away at Stanford or MIT,
both, not that I'm advocating for that, obviously, like we're still running on this trend of we're understanding more about life and we're getting better at tinkering with it.
Those trends will continue for the rest of our lives.
We had a really great talk with Drew. You know, I think the eye-opening thing for me from listening and
chatting with Drew in our podcast and the other conversations that I've had with him is that he
really does think like both a scientist and an engineer. So he is trying to turn many things about biology, which is incredibly tricky, into engineering platforms that we can then use to do in the biological sciences historically. So it is a
very, very radical shift in the way that we think about these disciplines.
Honestly, it blows my mind, and I love it. Here's a bit more from your conversation with Drew Indy.
Yeah, I think that was one of the quotes or things that I took away from our first meeting,
the fact that we don't completely understand a single cell in the human body.
Or any cell, or any cell at all.
Even the simplest microbe, there's not a single microorganism on Earth we understand completely.
Yeah, and we're sort of tangibly wrestling with this right now.
You've got this SARS coronavirus too, this little,
you know, 50, 100 nanometer particle that is like really doing a number on civilization right now.
And, you know, like I'm sort of glad that it's happening now versus 30 years ago, because we
have, as a matter of fact, come a very long way in our understanding of these biological systems over the past several decades.
But still, I think we're in many ways completely flummoxed by the mechanism of this virus and why it does one thing to one person and another thing to another. And like, even when you get down to the, you know, we sort of got lucky,
you know, as you mentioned in that first meeting that we had a solved structure for the spike
glycoprotein pretty quickly, you know, in the outbreak. And I know a bunch of work that people
have done to simulate in computers the interaction of that spike protein with these
ACE2 receptors in the human body, which is the mechanism that the virus used to invade a cell.
But even those computer simulations are relatively low resolution compared to what the actual
in vivo interactions are of that virus spike protein in the cell.
So we do have a long way to go still.
Yeah.
And honestly, we're playing.
We're not serious about biology yet.
We're not treating biology like a strategic domain.
When an envelope RNA virus can take out a carrier task force, something that no number of Chinese submarines can do apparently.
And all we can do is do F-15 flyovers to celebrate the healthcare workers.
That means we are not taking biology seriously. We are misspending our treasure.
30 years ago, by the way, 30, 40 years ago, by the way, it was HIV. And we had that experience.
So here's a question I'm wrestling with.
Why in infectious disease and epidemiology is it okay for us to adopt a strategic posture of let's wait till we're surprised?
Like, I don't know of any important strategic domain where, you know, community gets together or the leaders get together and say, well, we were really worried about this issue. And so our strategy is going to be,
we'll wait for something to happen and then we'll react and let's get better at reacting.
Like that's bizarre. And I think it's linked back to biology happens to us.
So that was Stanford professor and bioengineer Dr. Drew Indy.
And next we'll hear from neuroscientist and bioengineer Dr. Tom Daniel.
Yeah, Tom is a professor and faculty member at the University of Washington.
His research and teaching meld neuroscience, engineering, computing, and biomechanics to understand the control and dynamics of movement in biology.
So, if I remember correctly, Kevin, you and Tom spend a good amount of time talking about flies,
like this is flies. Yep. Tom does this really fascinating work with exploring how insects
navigate the world by looking at their neurobiology and electromechanical systems. And he told us all about these things on flies called haltiers.
According to Tom, a haltier is derived from the fly's hind wing, and they are these super
tiny non-aerodynamic knobs that flap just like wings.
Just completely fascinating.
So I will let Tom explain it.
He can do a much better job than I can.
They're like tiny dumbbells that the fly oscillates, a counter phase to the wings.
They're so small, there's no aerodynamic forces, but they're packed.
They're festooned with sensory structures.
As it turns out, they're like this little knob on a stick, and as that vibrates, it experiences bending forces. But if the fly rotates in a direction
orthogonal to the flap, it generates a Coriolis force, a gyroscopic sensor. And lo and behold, these systems are exquisitely sensitive
to rotational forces. So they're basically measuring, I apologize for the math,
the cross product of their flap with their body rotation. Okay. And so we had this idea that they're able or physically able to respond to Coriolis forces, but we really wanted to nail whether the neural system actually has the equipment to measure that.
And so we were able to stick electrodes into the neurons that go into these tiny modified hind wings
and measure their encoding properties.
And you can show that they encode information
at astronomically high rates
and do so for Coriolis forces.
That sort of led to an interesting question is,
these are what you would call a vibrating structural gyroscope, which is basically the
same idea that you have in all these gyroscopic sensors in your cell phone or anything else.
They operate at a tiny, tiny fraction of the energy cost. I'm not going to stick a fly inside
my cell phone, but bear with me, we do some odd things like that.
It's so wonderful and weird. I really, really like it.
And here's another bit from later in Kevin's conversation with neuroscientist Tom Daniel.
You have this beautiful point of view because you've been doing this for a while.
So what are some of the interesting things that have changedel the field much further forward than I will
see in my career. Candidly, ML methods, machine learning, is coming to bear on a vast number
of problems in neuroscience. Everything from imaging to how do we handle the massive data
flowing in from neural systems? How does a brain handle massive data? Can ML give us some insight?
So as we said not too long ago, there's lots and lots of channels coming in.
That's a hard problem to do in traditional control theoretic approaches, right? This is hard. And by the way, they're nonlinear.
You know, ML methods, I think the advent of AI and ML and our ability to grapple with massive data
is transforming the field of neuroscience, period. It's transforming the field of movement control.
We have the same problem in understanding how multiple actuators operate a dynamical system,
and how billions of motor molecules conspire to create movement in muscle.
These are all problems that demand extreme advances in computation,
not just the hardware of computation, but the ML methods that are coming about.
So even at my late stage of career, I'm finding myself having to learn more and more ML methods.
This is great. This is exciting.
So DNNs, even simple, just standard classification problems are becoming increasingly important.
That's revolution one that's been going on.
Revolution two is, of course, the advances in device technologies. So an example
of that will be the microfabrication of electrodes that you can implant in neural systems that record
from hundreds of simultaneous sites. I almost said thousand because it's at about 900 and something i think on the latest sharp electrode developed
for mouse brain recordings okay those are now device technology and of course the ubiquity
of micro fabrication is influencing even how we make electrodes interfacing with natural systems
okay so now you have these two things. You have ML methods, device technologies,
hand-in-hand, transforming our ability to understand the encoding and decoding processes
of natural systems. So what's the third revolution? Third revolution, of course, is gene editing.
Where is gene editing coming into all of this? Well, our ability to look at neural circuits
depends on our ability to look at variants in these neural circuits,
to turn them on, to turn them off,
to use optogenetic methods, to use CRISPR,
to change the chemosensory pathway on the antenna of an insect
with really awesome electrodes inserted into it
and ML methods listening in.
So those are the three technologies I think are transforming
not just neuroscience, I think it's transforming,
they're all mutually transforming each other.
That is, as we need to grapple with ever more complex data sets,
I think that's driving development of ML.
I think it's driving how we manage and control
and handle rapid information flow.
Just like real brains,
computers are faced with this real-time challenge.
Even the brain the size of a sesame seed
does astronomical amounts of computing at tiny levels of efficiency.
So there's lessons to be learned both ways.
You can tell I'm really excited because I see these synergies and this sort of triumvirate of advances in gene editing, advances in device technology, and advances in ML.
So that was University of Washington professor Dr. Tom Daniel.
Next, you'll hear some highlights from a recent episode with Dr. Oren Etzioni.
Oren is chief executive officer at the Allen Institute for Artificial Intelligence.
Yeah, I talked with Drew and Tom about the complexities of understanding biology. And with Orem, we got into the feasibility of creating AGI, artificial general intelligence.
I think it's a really interesting time to be a computer scientist, to be a computer professional. I do want to say, off the top of my head, here are three things
that the current technology doesn't
yet touch. The first one is
the current technology, maybe this is a good phrase, is kind of profligate
in its use of compute and data. Yeah, I need
millions of examples, at least for pre-training,
and then thousands for tuning.
Yeah, I need this massive amount of computation,
millions of dollars of computation to build my model.
And whereas, of course, human intelligence, which is the standard,
sits in this little box that's on top of my neck and is powered by the occasional salad and a cup of coffee, right?
So we're – and we know, right, you know, kids, they'll see one example and they'll go to the races.
So I think we can build far more frugal machines in terms of data and compute.
That's one.
And then the second thing, and this goes right back to the discussions we were having at CMU in the early 90s, is what is the cognitive architecture?
In other words, okay, you can take a narrow question like, is this email spam or not?
Or did I just say B or P, right?
Speech phoneme recognition.
And you can train models that'll do,
they'll have superhuman performance of that.
But the key thing in artificial general intelligence
in AGI is the G.
So how do we build what was called then
a unified cognitive architecture?
How do we build something that can really move fluidly from one
task to another when you form a goal, automatically go and say, okay, here's a sub-goal. Here's
something I need to do or learn in order to achieve my goal. There's just so much more to
general intelligence than these savant-like tasks that AI is performing today. The third topic in AI that I think we ought to be
paying more attention to is the notion of a unified cognitive architecture. So this is
something we studied at CMU back in the day, and it's the notion of not just being a savant, not just taking one narrow
problem, but going from one problem to the next and being able to fluidly manage living where
right now we're talking. Soon I will be crossing the street, then I'll be reading something.
Putting all those pieces together and doing it in a reasonable way is something that's way beyond the capabilities of AI today.
Yeah, and we've got a little bit of that starting in transfer learning, but just beginning.
Right, but the thing about the transfer learning is that it's still from one narrow task to another.
Maybe it's from one genre of text to
another genre of text. We don't really have transfer learning from, okay, I'm reading a book
to now I can take what I read in the book and apply it to my basketball game, right? We're very
far from anything like that.
Orin and I also talked about the use of AI
as augmented intelligence.
I asked Orin what he thought we should be thinking about
to better be prepared for all the innovation
coming in the near-term future.
Well, in terms of policy,
I think we do actually have to be very careful
not to use the kind of blunt and slow and easily distorted instrument of regulation to harm the field.
So I would be very hesitant, for example, to regulate basic research.
And I would instead look at specific applications and ask, okay, if we're putting AI into vehicles, how do we make sure that it's safe for people?
Or if we put AI into toys, how do we make sure that's appropriate for our kids?
For example, the AI doesn't elicit confidential information from our kids or manipulate them
in various ways.
So I'm a big believer in regulate the applications of AI, not the field on its own. I think some of the
overarching regulatory ideas, for example, in the EU, there's the right to an explanation.
And it sounds good, right? AI is opaque. It's confusing. These are called black box models.
Surely, if an AI system gives us a conclusion, we have a right to an explanation. That sounds
very appealing. But I actually think it's a lot trickier than that because there are really two kinds of
explanations of AI models. One is explanations that are simple and understandable but turn out
not to be accurate. They're not high-fidelity explanations because the system is complex.
And a great example of that is if you go to Netflix and it recommends a movie to you,
they've realized that people want to know, why did you recommend this movie? And say, well,
we recommended this movie because you liked that movie, right? We recommended Goodfellas because
you liked The Godfather. Well, if you look under the hood, right, the model that they use is
actually a lot more complicated than that. So they gave me a really simple explanation. It's
just not true.
So that's one kind. The other kind is I can give you a true explanation, but it'll be completely incomprehensible. So now if the EU says, you know, you have a right to an explanation, what you're
going to end up with is one of these two horns of the dilemma, something that's incomprehensible
or something that is inaccurate. So I think that
it's really important that we are careful not to go with kind of popular notions like right
to explain, but instead think through what happens in particular contexts.
Yeah, I think that is an extraordinarily good point. These models are already at the complexity
where they're as complex as some natural phenomenon.
We're not able to explain many natural phenomena,
you know, because, you know,
when we get down to the point of like,
these are the electrostatic interactions of the atoms
that comprise this system,
you have to look at the phenomenology of the system. It's why statistics
is going to be such a really important skill for everyone. It's why understanding the scientific
method and having an experimental mindset, I think, is important. I think this is such a good
point about not deceiving ourselves that an incomprehensibly complex answer
to a question of why did this thing do what it did,
even if it's couched in terms of language that we might otherwise understand,
that's not real understanding.
Exactly. And I'm not suggesting that the solution is,
hey, just trust us, we're well-intentioned AI people. It's going to work.
But again, going back to the auditing idea,
rather than an explanation,
if we want, you know, one of the most jarring ones
are uses of AI in the criminal justice system, right,
to help make parole decisions and things like that.
Well, we should audit these systems, test them for bias, right?
The press should be doing that.
The ACLU should be doing that. Reg. The press should be doing that. The ACLU should be doing
that. Regulatory agencies should be doing that. But the solution is not to get some strange
explanation for the machine. The solution is to be able to audit its behavior statistically
and test it. Hey, are you exhibiting some kind of demographic bias?
Yeah. I mean, one of the things that we do at Microsoft is we have these two
bodies inside of the company, this thing called the Office for Responsible AI that sits in our
legal team. And we have this thing called Ether that's the AI and Ethics Committee inside of the
company. What we do with both of these bodies is we try to have both the lawyers and the scientists thinking about
how you inspect both the artifacts that you're building in your AI research, but their uses.
And we have a very clearly defined notion of a sensitive use. And depending on how sensitive a
use a particular model is being deployed in, we have different standards of auditing and scrutiny that go along with it and recommendations.
Like for a criminal justice application, for instance, you may say that a model can only advise.
We do not condone it making a final decision, you know, just so that there's always human review in the loop.
I think that's smart. And I also think that this relates to another key principle when we think
about both regulatory frameworks and ethical issues. Whose responsibility is it? The responsibility
and liability has to ultimately rest with a person. You can't say, hey, you know, look, my car ran you over. It's
an AI car. I don't know what it did. It's not my fault, right? You as the driver, or maybe it's
the manufacturer if there's some malfunction, but people have to be responsible for the behavior of
the machines. The same way that, look, the car is already a complex machine with 150 CPUs and so on.
I can't say, oh, well, the car ran you over.
I had very little to do with it.
The same is true when I have an AI system.
I have to be the one who's responsible for an ethical decision.
So I very much agree with you there.
That was the CEO of Allen Institute for AI, Dr. Oren Etzioni.
Now, here's a part of the show that I've most been looking forward to because I get to ask Kevin some questions.
But before that, let's hear from this past year's episode recorded right before the launch of Kevin's book, Reprogramming the American Dream from Rural
America to Silicon Valley, Making AI Serve Us All. Kevin got together with his co-author,
Greg Shaw, to talk about the book. Greg is a former journalist and has worked with Bill and
Melinda Gates, both at Microsoft and at the Bill and Melinda Gates Foundation. He is also co-author
of Satya Nadella's book, Hit Refresh. We'll start with Kevin reading an excerpt from the book.
With modern technology, with more of our time spent online and on our devices, and with
more and more of our connections with one another mediated by social networks, it's
hard to avoid becoming trapped in self-reinforcing filter bubbles
and then not to have those bubbles exert their influence on other parts of our lives.
Many of my friends and colleagues see those living in rural communities, people who live
outside of the urban innovation centers where the economic engines are thrumming right now,
in a very different light than I do. That's not just unfortunate. It's an impediment to making
the American dream real for everyone. The folks I know in rural America are some of the hardest working,
most entrepreneurial, cleverest folks around. They can do anything they set their minds to
and have the same hopes for their futures and the futures of their families and communities
as those of us who live in Silicon Valley and other urban innovation centers all do.
They want their careers
and their families to flourish, just like everyone else. Where we choose to live shouldn't become a
dividing line, an impediment to a good job and a promising future. That's the American dream,
and it's on all of us to make sure that it works, because in a certain, very real sense,
if it doesn't work for all of us, it won't work for any of us.
Fantastic. You know, Kevin, now that the book has been out there for a few months,
I wanted to ask you, how's the book been received?
I think it's been received pretty well. I mean, it's obviously an unusual time to launch a book and people have an awful lot of things to think about. But the reason that I wrote
the book in the first place was to try to help enrich the conversation that we were having about
artificial intelligence, how it could be used and how it shouldn't be used. And I think it has done
a better job than even I was hoping for in forming those conversations. And it's certainly
been wonderful to have conversations around the book with people I never would have had the
opportunity to chat with before. You know, Kevin, on this podcast, you've shown a lot of curiosity
about our guest origin stories. You know, where do they come from? How do they get started in their field?
And in your book, you talk about your grandfather, Shorty, who is this remarkable character.
Why did you feel that it was important to include these anecdotes about your grandfather, your mom, your dad?
Why is it important for today's leaders and innovators to share parts of their origin story?
Well, I think there are a couple of reasons. One of the things that I wrote in the book is I really do believe that we are the stories
that we tell.
Storytelling really does shape who we become.
They sort of give us a roadmap for the future that we aspire to have for ourselves and for
each other. And I think it's really important for all of us to sort of share our stories because, you know, there are interesting lessons in there. But maybe most importantly, understanding each other a little bit better shows us how I do honestly believe that we are vastly, vastly more similar than we are different.
And that sometimes these stereotypes that we have of each other are just obstacles that stand in the way of each of us achieving our full potential.
Yeah.
Yeah, I like that.
I like that.
Here's more from Kevin and Greg's conversation.
You dedicate the book to your father.
In the book, you write a letter to your grandfather, Shorty explaining to him he was obviously a craftsman and someone who would have been fascinated by AI. I mention this because the book is titled Reprogramming the American Dream, and you had your family and other families in mind.
What's involved in reprogramming the American dream, and what do you mean by the American dream? an opportunity with better investment in advanced technology and like making those investments in a
way where they're you know accessible to as many people as humanly possible to have people in rural
and middle america have the opportunity to create really very interesting new businesses that create jobs and economic opportunity
and that help them realize their creative vision and that, you know, serves as a platform
in the same way that industrial technology has served as a platform for these communities
to build their economies in the, you know, in the early mid 20th century, that AI can have a similar sort of
effect in these communities today. You offer a number of different suggestions related to
education and skilling and that sort of thing. I'm curious, what would you say is your advice
to young people who might be growing up in rural central Virginia or Oklahoma, where I'm from?
How should they prepare
for jobs of the future? Yeah, I've chatted with a bunch of people about this over the past
few weeks. And, you know, when I get this question about what we need to do to make AI accessible to
those kids in rural and middle America, yeah, some of the things that we need to do are just very
prosaic, I think. So the tools
themselves have never been more powerful. Like the really interesting thing to me is that first
machine learning project that I did 16 years ago now required me to sit down with a couple of
graduate level statistical machine learning textbooks and a whole stack full of fairly complicated research
papers. And then I spent six months writing a bunch of code from scratch to use machine learning
to solve the particular problem I was trying to solve at the time. If I look at the state of
open source software and cloud platforms and just the online training materials that are
available for free to everyone, a motivated high school student could do that same project that I
did 16 years ago, probably in a weekend using modern tools. And so, you know, I think that the
thing that we really need to be doing is figuring out how to take these tools that are now very accessible and like we shouldn't we shouldn't feel intimidated by them in any shape, form or fashion and figure out how to get those into high school curricula so that we are teaching kids in a project oriented way, like how to use these tools to solve real
world problems. I think getting kids those skills is super important. Like the other thing that we
need to think about is just how we're connecting people to the digital infrastructure that is
going to increasingly be running our future.
And so there are things like the availability of broadband that are a huge, huge deal.
I think we write about in the book, my visit to our data center in Boynton,
which is in Mecklenburg County, about an hour and a half, two hours away from where I grew up.
And this is one of the most sophisticated technology installations anywhere in the world.
Like there's an enormous amount of network bandwidth coming into this facility and like
the amount of compute power that is just in this sort of acres of data center infrastructure
that we have there is just staggering.
And we have a bunch of high-skill technology workers who are building and operating this
infrastructure on behalf of all of Microsoft's cloud customers.
And some of those people who are living in that community struggle to get access from
their local telecommunications
providers to the high-speed broadband that they expect.
Like, they're information workers.
Like, they expect in their homes to, like, have good broadband connectivity.
For students, like, it's even more critical.
Like, if you don't have a good broadband connection that's available to you somewhere as a student.
Like, you're never going to be able to go find these open source tools to use these
free or cheap cloud platforms to, like, go learn all of this, like, very accessible knowledge
that is on YouTube.
And so sometimes I think it's the, you know, the prosaic things that, like, we're making
more complicated than the complicated things.
That was Kevin speaking with his co-author, Greg Shaw, about their latest book, Reprogramming the American Dream.
Now let's switch gears a bit and meet one of Kevin's favorite science fiction authors, Charlie Strauss.
Kevin, tell us a bit about Charlie and why you invited him on the show.
Well, I obviously love Charlie's fiction.
He's a phenomenal writer.
And, you know, I feel like I've been really, really shaped by the books that I've read
from early in childhood.
And I've been, I think, especially inspired and motivated by the science
fiction that I've read. And Charlie's is certainly a really, really wonderful, wonderful fiction.
I first found out about Charlie as a writer when I read his books on the singularity,
which is this sort of notion that computers become intelligent and self-evolving and so rapidly self-evolve that they become this sort of unknowable thing and weirdness happens, right?
Like once you cross over the singularity, things become very unpredictable. And so having someone with an amazing, brilliant imagination, and in Charlie's case, an actual background in the technology industry because he was a programmer and a technical writer, really can shine a light on these sort of crazy circumstances that we can set up for ourselves as acts of imagination. And so I just thought it would be really interesting to hear Charlie's take on the current state
of affairs in the world where technology has become an increasingly important part of all
of our lives and an increasingly important factor shaping what's going to be happening
in the future.
And as someone whose job it is
to literally imagine the future,
I thought he might have something interesting to say.
For sure, for sure.
Well, let's hear from Kevin's conversation
with science fiction writer
and Hugo Award winner, Charlie Strauss.
How do you start this process
of trying to imagine what the future might be like so that you can have a foundation for the stories that you're telling?
Okay. I don't always start from the perspective of the world building itself.
I usually start from a point of view of the characters because fiction is essentially the study of a human condition under circumstances that don't currently apply. And, you know, if you're going to talk about the human condition,
you have to start by talking about people. Having said that, there are a couple of books I wrote
in 2006 and 2009, which were very tightly focused on the world 10 years in the future.
It was going to be a trilogy,
but unfortunately, the third book in the trilogy has been persistently derailed by political developments in the real world. I mean, I just can't write it. I've had about two or three
different plots for it, both destroyed. The most recent one, it was killed by COVID-19,
because I do not want to write a book about a viral pandemic at this point.
Yeah. viral pandemic at this point. Anyway, those books were Halting State and Rule 34. And the idea of
Halting State, I got in 2005 when I was at a science fiction convention at a panel discussion
discussing massively multiplayer online role-playing games like world of warcraft at that point and a member of a panel
who are on the top table at that point came up with a couple of points the first was that mmos
were the first commercially successful virtual reality environment one in which you have lots
of people with avatars meeting each other um forget the lack of headsets or tactile feedback or head positioning
and so on. There's still a window into a virtual world. The second thing he came up with was
there's economics involved. He gave as an example an anecdote of an incident that happened in London
a couple of years earlier when a guy walked into a police
station to report a crime. Somebody he'd met on the internet had sold him a magic sword,
and it wasn't magic. That's great. It turned out to be fraud. He bought a weapon inside a game
via an eBay auction, and it wasn't as described. It did actually get written up as a fraud.
And I suddenly realized at this point, hang on, I need to do some digging here.
And I did some digging and discovered some exotic studies,
including one paper that confirmed if you take in-game currencies
and convert them to real-world currencies using whatever players are running as an exchange rate by about 1999 there
was one game which had an economy with about the same value as the gdp of austria but well no you
can't really do a real world conversion like that because it's just fatuous you'd crash the in-game
economy if you tried anyway but uh there was something going on here and you know economics is the study of how human beings
allocate resources under conditions of scarcity and i began to ask myself what's the world going
to look like in 10 years time if we really do get artificial augmented reality goggles and
self-driving cars and computer games everywhere and mmos and and live-action role-playing combined with
high-bandwidth always-on stuff. So I started designing what I thought the world of 2017
would look like. And I got this book written and had a bit of a hard time selling it in the UK,
although it sold well enough in the US. The problem is a crime novel set in 2017 in an independent Scotland where it opens with a cop being summoned
to the boardroom of a startup company she doesn't understand in a former converted nuclear bunker.
To be told there'd been a bank robbery. A gang of orcs with a dragon for fire support had robbed
the central bank inside an Mmo and she gradually well various
consultants are called in including forensic accountants and a computer guy because you need
a computer guy for this sort of stuff and so on and it we gradually discover that somebody has
come up with a exploit for compromising the private keys of a company whose basic speciality is arbitrage
between the economics of competing MMOs, because one game's company after another is trying to
poach their competitors' customers. And there's now a capture of a flag game in progress between
rival teams of Chinese hackers who are trying to hijack the economy of a small European state.
Now, to get to where this was going to go, I tried to do some rigorous extrapolation,
came up with a couple of rules of thumb. And the first is, if you're looking 10 years in the future,
70% of that world is here today. About half the cars on the street, they're already there. You know, they're going to be there in 10 years' time. They're still going to be driving. They're going to be a bit more decrepit, but they're out there. Buildings, the average house
in the UK is 75 years old. I know American dwellings tend to be a lot younger, but 10 years'
time, there's not going to be much turnover. There'll be a few new office buildings, a few new
developments, but most of what we see is there. The people, everybody's going to be 10 years older.
The people at the top of the age range will, well, they won't be visible anymore. The kids,
they're going to be teenagers, but it's the same stuff. 70% of it is there today.
You then get another 20%. No, actually, it's about 80% that's there today. You then get about another 15% that is pretty much predictable.
It's on roadmaps.
We knew back in 2006, 2007, that by 2017, we'd have 3G cellular telephony as standard,
and something called 4G would almost certainly be out there, but not universal by then.
I had no idea what the 4G standards were, but 3G was pretty much visible. Everything back then was running on GSM.
The state of the phones we were using, again, it was fairly obvious that they would be connected
devices and they'd be very smart pocket computers. I missed a call on that by going for artificial
reality goggles, shades of Google Glass, which, as we know, kind of crashed in the market for social reasons rather than technological feasibility.
It may eventually happen.
There's always, though, an element of a couple of percent, which is who ordered that?
You know, stuff that comes out of left field completely, and it's completely unpredicted. The CCD image sensor that we have in all our cameras today was, I think, developed in the 1980s, actually commercialized.
And people realized that these things were literally cheap as chips.
What are we going to do with them? Where are we going to put them?
The idea that everybody would be carrying a decent quality camera around with them at all times, though video camera that could upload to the internet uh that was not something most people were prepared to grapple with and the idea that
there'd be a craze for happy slapping whereby teenagers would find a random stranger and video
one of their mates going up to them and beating the crap out of them and then put it on youtube
yeah luckily that was a short-lived craze most of the and then put it on YouTube. Yeah, luckily that was a
short-lived craze. Most of the people who did it didn't realize they were basically preparing
evidence for prosecution. But it's a second-order consequence. As Frederick Pohl once said,
anyone can predict the automobile. The difficult bit is predicting the traffic jam.
That was science fiction writer Charlie Strauss.
Yeah, and don't forget to grab a copy of Charlie's new book, Deadline Streaming. It just came out in October, and it is fantastic. As mentioned, I'm intrigued to learn about those first sparks of
curiosity that lead our guests into their professional pursuits. There's so many common threads between them.
One of my favorite stories this year was from Microsoft's own Eric Horvitz.
Yeah, Eric is a Microsoft Technical Fellow and our very first Chief Scientific Officer.
And he provides expertise on a broad range of scientific and technical areas,
from AI and biology and medicine to a whole host of issues
that lie at the intersection of technology, people, and society. Let's hear a bit from
Kevin's conversation with Eric. So I'd love to start, as we always do, by understanding how it
is you first got interested in science and technology.
Presumably that was when you were a kid. So can you tell us a little bit about that?
Yeah, I just know that I've always been sort of inspired to understand things,
and I didn't distinguish between human creations, artifacts, and stuff I would see in the world.
So I was confused and intrigued and interested in living things in space and time.
I remember even being very, very young, asking my first grade teacher if I can know more about time.
And she ended up bringing me to the library at Birch Elementary School and showing me a book about clocks.
And I said, no, showing me a book about clocks.
And I said, no, I don't really mean clocks.
I mean time.
And I was also intrigued by light. I had this really beautiful phosphorescent, phospholuminescent nightlight in the 60s.
And beautiful green light would wash the room at night in this glow.
And I was just curious, what the heck was light?
So I had these basic questions I remember having a discourse with my father about um you know I heard a lot about God
and I was curious what God was made of and I couldn't get a good answer from adults about that
and when it comes to machines and mechanism I took apart a flashlight I think it was like the
summer after kindergarten or so,
because I remember in first grade I was already into this and talking to friends about this.
But I realized that there was a circuit there, and I found some wire,
and I think I impressed my family more than myself when I ran around the house with a battery and a wire
with a light bulb lighting up with my finger, under my finger. And I think this was also around the time that, again, mid-60s,
when there was a lot of the cartoons we were watching back then
had electronic robots and Astro Boy flying around,
very helpful entities.
And I was curious about electronic brains.
I don't know where I got that idea.
But I remember having a bag of parts
on my way to my grandmother's house
in the back of a station wagon.
Maybe this is around second or third grade.
But the peanut can, wires, light bulbs,
I thought I could assemble an electronic brain
on the way to my grandmother's
house in the back of a station wagon and didn't get, you know, still working on that today,
basically. That's really awesome. And were your parents scientists or technical engineers?
My parents were both school teachers. My mother was a kindergarten teacher. And I remember being
very proud of that in kindergarten. I would tell everybody at a mother was a kindergarten teacher, and I remember being very proud of that
in kindergarten. I would tell everybody at a time when the kindergarten teacher was like the person
you most looked up to that, by the way, my mom was a kindergarten teacher too. That was considered
awesome by my peers at the time. And my father was a high school teacher. He did science as well as history.
So where, I mean, it sounds like you had a bunch of innate curiosity, which is awesome.
And like one of the themes I think we see with a lot of people who chose careers in science and technology.
But did you have any role models when you were a little kid or things that were in the popular media that were inspiring you? Or did this just really come out of,
you know, from your perspective, nowhere? Lots of books. My parents had a home library
filled with lots of books. We had the Merrick Library, Merrick, Long Island, where I would
spend lots of time. I got to know the science sections as well as the pet section of the library pretty intensively.
And mostly books at the time and friends, some of whom had aligned interests.
It's hard to think of the idea of being in first or second grade
and having a third grade having a scientific support team.
But we sort of had peers that were interested as well.
In third grade, I became, I was elected to be the chairperson of the science club, I
remember.
And we had all sorts of projects involving wind speed and solar energy back in those
days.
But I'm not sure, you know, where some of the interest came from.
It was
largely curiosity and books. And later in life, of course, I had some fabulous mentors. And,
you know, we all think back to our various teachers in elementary school. You know,
you start in kindergarten and go to sixth grade. Each teacher has a major influence on people.
And, you know, I can remember sitting at this desk
in sort of what I thought was kind of a militaristic setting,
and I asked myself on the first day of first grade,
is this what school's going to be like?
I have to sit at this desk for like 12 years.
And the way that the first grade went,
I was really unimpressed,
and I would have given it all up if it wasn't for, and I'll call out a name, Mrs. Frank, my second grade teacher, who like completely opened the world to me.
Was open to science and interested in answering questions.
You know, and then you jump forward to fifth grade, Mrs. O'Hara.
And these people were just brilliant teachers.
Mr. Wilmot in sixth grade, he celebrated my interests.
And we had science fairs and actually won the science fair that year.
And you have a few teachers like that that really are like large planets that spin you off with the gravitational field into new directions.
That was Dr. Eric Horvitz, Chief Scientific Officer at Microsoft. And I really
recommend listening to the entire podcast. It was a fascinating conversation that delved into,
among many other things, the intersection of biology, AI, and high-performance computing
that's been one of the themes this year. Yeah, it was a great episode. They all are,
if I do say so myself.
And, you know, in that clip that we just heard, I really loved that Eric gave a shout out to the teachers who inspired him and helped him become who he is today. So in the spirit of celebrating
all of our invaluable educators out there, I would like to give a shout out to Ms. Cohen.
Kevin, what about you? Oh, that is an extremely difficult question, just to name one.
I had so many teachers who had such a phenomenal impact on me.
Maybe I'll give a shout out to Dr. Tom Morgan, who taught me my first real bits of computer science when I was in high school at the Central Virginia Governor's
School. Shout out to Tom. That's awesome. Well, the show would not be complete if we did not hear
from at least one of our guests about what they do in their spare time. So to wrap things up,
we'll hear from Dr. Percy Lang. Percy is an associate professor of computer science at
Stanford University and one of the great minds in AI, specifically in machine
learning and natural language processing. One last question. So, just curious what you like
to do outside of work. I understand that you are a classical pianist, which is very cool.
Yeah. So, piano has been something that's always been with me since I was a young
boy. And I think it's also been a kind of a counterbalance to all the other kind of tech
heavy activities that I've been doing. What's your favorite bit of repertoire?
I like many things, but late Beethoven is something I really enjoy. I think this is where he becomes kind of very reflective about,
and his music has a kind of an inner, it's very kind of deep.
And so I kind of enjoy that.
Like what particular piece is your favorite?
So he has a Beethoven sonata.
So I've played the last three Beethoven sonatas, Op. 109, 110, 111.
They're wonderful pieces.
Yeah.
And one of the things that I actually, you know,
one of the challenges has been incredibly hard to make time for a serious hobby.
And actually in graduate school,
there was a period of time when I was really trying to
enter this competition and see how well I could do.
Which competition?
It was called the International Russian Music
Piano Competition.
It was in San Jose.
I don't know why they had this name.
But then, you know, I practiced a lot.
There's some days I practice like eight hours a day.
But at the end, I was just like, this is, it's just too hard.
I can't compete with all these people who are kind of the professionals.
And then I kind of, I was thinking about how,
what is the bottleneck?
Often I have these musical ideas
and I know what it should sound like,
but you have to do the hard work
of actually doing the practicing.
And, you know, kind of thinking maybe wistfully,
maybe machine learning and AI
could actually help me in this
endeavor. Because I think it's kind of an analogous problem to the idea of having a desire and having
a program being synthesized or an assistant doing something for you. I have a musical idea. How can
computers be a useful tool to augment my inability to find time to practice?
Yeah.
And I think we are going to have a world where computers and, like, machine learning in particular, like one of those disciplines, and like there are several of them, where it's just blindingly obvious that the difference between expertise and non-expertise, like no matter how much I understand.
And so like I'm not a classical pianist.
Like I'm just an enormous fan.
Even though I understand harmony, I understand music theory, I can read sheet music, I can understand all of these things, and I can appreciate Martha Argerich playing, you know, Liszt's Piano Concerto No sit down at the piano and do what she does because she has put in an obscene amount of work training her neuromuscular system to be able to play and then to just have years and years and years of thinking about how she turns notes on paper to something that communicates a feeling to her audience. And it's like really just to me stunning because there's just no,
there's no shortcutting it.
Like you can't cheat.
Yeah.
It's kind of interesting because in computer science,
there's sometimes an equivalence between the ability to generate and
ability to kind of discriminate and classify, right?
If you can recognize something, whether something's good or bad,
you can use that as objective function to hill climb.
But it seems like in music, we're not at the stage where we have that equivalence.
I can recognize when something is good or bad, but I don't have the means of producing it.
And some of that is physical, but I don't know.
Maybe this is something that is in the back of my mind, in the back pocket,
and I think it's something that maybe in a decade or so I'll revisit.
The other thing, too, that I really do wonder about with performance is there's just something about, like for me it just happens to be classical music.
I know other people have these sorts of emotional reactions
to rock or jazz or country music or whatever it is that they listen to.
But I can listen to the right performance of like Chopin's G minor ballad.
And like there are people who can play it and like I'm like,
oh, this is very nice and like I can appreciate this.
And there are some people who can play it,
and every time I listen to it, 100% of the time, I get goosebumps on my spine.
It provokes a very intense emotional reaction.
And I just wonder whether part of that is because I know that there's this person
on the other end, and they're in some sort of emotional state playing it
that resonates with mine,
and whether or not you'll ever have a computer
be able to do that.
Yeah, I mean, this gets kind of a philosophical question
at some point.
If you didn't know it was a human or a computer,
then what kind of effect would it have?
Yeah, and I actually had a philosophy professor
in undergrad who asked the question, would it make you any less appreciative of a Chopin composition knowing that he was being insincere when he was proposing it?
He was doing it for some flippant reason, and I was like, yeah, I don't know.
Well, one of my piano teachers used to say that you kind of have to, it's a kind of like a theater.
You have to convey your emotions,
but there has to be some, even when you go wild,
there has to be some element of control in the back
because you need to kind of continue the thread.
Yeah, for sure.
Yeah, but also it is, for me also, just the act of playing it is a pleasure.
It's not just having a recording that sounds good to me.
Yeah, I know I'm very jealous that you had the discipline and did all the hard work to put this power into your fingers.
It's awesome.
Well, I think that's a great note to end on.
Did I detect a little pun there?
Yes. Yes, you did, Kevin.
Well, before we close, I just want to say thank you again to our guests this past year.
I'm so grateful to all of the folks who shared their time and their vision with us.
Yes.
And as always, thank you for listening.
As 2020 draws to a close and a fond farewell, might I add, please take a minute to drop us a note to BehindTheTech at Microsoft.com and tell us about your hopes for 2021.
Be well.
See you next time.