a16z Podcast - Pandemics: Early Detection, Networks, Spreaders
Episode Date: May 23, 2020Pandemics are predictable; what's not predictable is the intensity, or the precise timing of arrival. That's where early detection -- not just rapid warning (as with something like Google Flu Trends b...ack in the day), or even delayed warnings (as with CDC flu trackers and such) -- comes in. Because unfortunately, many disease tracking efforts old and new are "like watching the weather forecast a week after you've experienced that weather", observes a16z general partner Jorge Conde.And this matters for saving lives; for load balancing and allocating resources (ventilators, PPE, supplies); getting back to work; and much more. Even a two-week advantage could have made a huge difference! Which is what sociologist and physician Nicholas Christakis (who directs the Human Nature Lab, part of the Yale Institute for Network Science, and also author of the book Blueprint) learned from the H1N1 pandemic. Specifically, the role of social network "sensors" -- where friends in one's network graph can be like canaries in the proverbial coal mine to help detect pandemics earlier.In fact, the lab recently released an app called Hunala (which uses information crowdsourced among networks) to determine one's likelihood of contracting flu/ influenza-like or other respiratory illnesses through a personalized daily assessment of risk. Kind of like Waze, but for illnesses not car accidents. So in this episode of the a16z Podcast, the two take that analogy far. They also discuss the role of other mobility data and population flows in China for where and when the pandemic spread; the nuances behind "superspreaders"; how bad is the coronavirus, really; and the near future of "bio-surveillance" -- not just from a personal risk perspective, but from a global public-health perspective... Can we get the holy grail here without sacrificing privacy and agency?
Transcript
Discussion (0)
Hi everyone. Welcome to the A6NZ podcast. I'm Sonal. Today we're continuing our ongoing podcast series covering different aspects of the coronavirus pandemic. As a reminder, you can catch all past and recent content on this at A6NZ.com slash coronavirus.
Today's topic is on tracking networks of spread for predicting pandemics and loads, which has implications for resource allocation, getting back to work, and much more.
Our guest is sociologist and physician Nicholas Christakis, who directs the human nature lab,
part of the Yale Institute for Network Science. He and his collaborators previously published
work in PLOS on the H1N1 pandemic, which informed their newly released app,
Hunala, which is an example of a privacy preserving approach for tracking pandemics to allow
not just rapid warning, but early detection. They also published a recent paper in nature on
how population flow drives spatiotemporal distribution of COVID-19 in China, end quote,
which among other things may comment on the accuracy of the early data released by Chinese authorities,
all of which is covered in this episode, and you can find all the links referenced in the show notes.
Also, a quick note that we have another episode on not just tracking and predicting,
but the privacy aspects of contact tracing, a separate but very related topic, on 16 minutes.
You can find that show in its own feed if you're not already subscribed there in your favorite podcast app.
In this episode, Christakis is interviewed by A6 and Z general partner Jorge Condé on our bioteam.
The two also share their thoughts on how bad is it, but they begin with where testing and tracking fits and what's been missing given that pandemics themselves are a lot more predictable than most people realize.
I need to emphasize just how predictable this is. It's not just, you know, for example, Bill Gates,
had this TED Talk in 2015, or Tony Fauci has been publishing papers about this for 20 or 30
years. I could reach over on my bookshelf right here and get a book called National Strategy
for Influenza Pandemic, and it describes everything we're experiencing. There's like a playbook,
like what you're supposed to do. What's not predictable is the intensity or the precise timing
of its arrival. So we were kind of lulled into a false sense of security. In 2003, we had the
first SARS epidemic, which ultimately only afflicted like 8,500 people were
worldwide. And it put the fear of God in Taiwan and Hong Kong and Korea and Japan and China, but not
us, because we had like 30 cases or something. And then in 2009, we had the H1N1, which did afflict more
individuals, but it was so mild that most people just had a cold. So again, we didn't take it
seriously. Well, now we have a perfect storm of a serious pathogen that's afflicting a large number
of people and were completely unprepared. And unfortunately, we weren't ready with the kind of testing we
needed to really respond to the epidemic. And just to be clear, testing has multiple purposes
and there are multiple different kinds of testing. One purpose is to identify people who need
treatment or identify people who need quarantine. But another purpose is an instrumentation function,
like you wouldn't fly a plane without knowing your altitude and your speed and so on. If you're
a policymaker and you're responsible for the health of the public, you kind of need to know where is
the germ and how much is there. And we were flying blind for weeks. And we still are honestly
flying blind. If you sort of break down the problem, it's because there are a bunch of
different things that have been missing in the past that I think we find ourselves having flown
blind here. Number one is we just need better detection technology. So here it'd be in the form
of testing. Number two, we actually need better mechanisms and frankly incentives for people
to share data more broadly, whether it's gathered at a governmental level or at institutional
level or ultimately at an individual level so that we can get sort of high resolution maps
of where disease is present and where it's not. And number three, we need the ability to
analyze these things in very real time. The single unifying goal of public health is to do this
and to do this well. One of the holy grails, I think, for public health has been to create
some comprehensive sort of biosurveillance, essentially a weather map of where there is health
and disease, you know, across the country or across the world. And specifically within the
U.S., you know, there are several efforts to sort of get at this. The CDC has a sort of flu tracking
map. I worked in a previous life on a product against a virus called RSV, respiratory-sinsidal
virus that affects primarily premature infants, elderly, and other immunocompromise people. And one of the
big challenges we had every single year, we had a therapeutic that would be given to premature infants
that were particularly at risk, was finding out where and when premature infants would be at
elevated risk, and therefore would be good candidates for prophylaxis treatment.
And what we found time and time again was that the resolution of the weather map, if you will,
was so low and so latent.
You know, it took, you know, we were usually getting reports in weeks and weeks later.
So there really wasn't sort of a way to get a good forecast of where to be with the appropriate
prophylactic treatment. And if we look over sort of the scope of the public health apparatus
around the world, and certainly in the United States, we're trying to come at this from two extremes,
I would say. On the one hand, we want to figure out, can we get accurate forecasts? And then the other
end of the spectrum is, how do we get to ground truth as to where the virus is? Well, on that issue,
we have this app called Hunala that is like ways for respiratory disease or ways for influenza
a like illness. So each person signs up, they contribute a little information, and then you can see
what is your risk of getting respiratory infections around the bend, like upcoming, the kind of
crowdsourced, non-governmental, privacy-respecting way of getting information about your risk of
coronavirus. The science behind it is stuff that we published for the H1N1 pandemic back in 2009,
And the gist is that if something is spreading in the network, it should reach central people,
people with many friends, whose friends in turn have many friends, sooner in the course of the
epidemic than it reaches peripheral people, people who have few or no friends, people who are
hermits. That means that certain people in the middle of the network are going to be more likely
to get anything that's spreading earlier in the course of the epidemic than random people.
As you probably know, about 10 years ago, Larry Brilliant's team proposed this notion of Google Fluet Trends, which was a kind of a big data idea of let's use what people are searching for today in order to know where is the epidemic today.
But if you could identify central people in a network, they would act as canaries in a coal mine because they're going to get the epidemic early.
If you can see the epidemic spiking in those individuals, you can know where the epidemic will be in two weeks, for example.
I like the Waze analogy because Waze has, you know, sort of two modes.
One is it gives you objective truth of where you are on the map and which direction you're heading.
And then it gives you, like, subjective truth and, like, who saw a cop and who ran over a pothole and all of those things.
And you want to ideally marry those two things.
I mean, yeah, there are two perspectives on trying to understand what's happening when it comes to the pandemic.
One is the individual perspective, like, I want to know what's happening around me, what's happening in my community, what's happening among my friends or the people I'm interacting with.
And then, of course, there's the collective perspective where public health authorities or hospital officials or politicians want to know what's happening.
Like, can we understand where is the germ?
We have a dashboard function, which would be as if the police or the state authorities interested in highway maintenance could monitor all these reports and see dozens of people are saying there's a pothole here or this is where accidents occur.
The same thing happens where hospital directors or public health authorities, if there's enough
adoption, that's the crucial thing of our app, it could also serve that function, like what's
happening right now in this district versus that district?
What's happening in my catchment area in terms of, you know, symptom reports, for instance?
So it's like if you can log on to Ways and see, today is a trafficy day in New York City,
but what about on my street, the part of the city that I am in, what can you tell me about
that. After you complete your symptom check-in for the day, you get immediate feedback on what is your
risk of infection given where you live and what is your risk of infection given where you are in
the social network. For example, your friends, friends, two weeks ago had the flu. And so now it's
winding its way towards you. When they first sign up, you also tell us a little bit about who your
friends are. And we use that data. It's all with respect to privacy. We don't track that information,
but we use that information to ping those people. And then the app uses machine learning to
process large amounts of information that you've told us about yourself and that the whole
network is generating to give you a personalized assessment of your risk of getting a respiratory
disease based on where you are. What is the assumption around network coverage in terms of
adoption that you would need for this to be informative?
So it is exactly like ways. The more people that use it, the better it is. The risk we run, of course, is that adoption will be low in any given location. If we have about 100 people in a county, we think that it will begin to perform much better than just naive feedback to the users. It's like trying to estimate how traffic he is New York City, one, two, three, four, five people telling you they're looking out their window and seeing the traffic isn't good. By the time you have 100 people,
people in the city saying, oh my God, traffic's at a stencil outside my window. Then you can
begin to make a generalized estimate. And the more people you get, of course, the more specific
you can get. And so to abuse the way's analogy for a second, how do you think about the potential
that this particular virus has a very, very high rate of asymptomatic people that either aren't
asymptomatic for a long time or remain asymptomatic over the course of the disease and therefore
or the equivalent of their invisible cars on the road causing traffic.
The analogy would be the cars that are not speeding
or cars that have not been pulled over or don't have any accidents.
So you're right.
We think about 50% of people are asymptomatic from this condition,
from coronavirus, the best estimates now.
Such individuals would still benefit, of course.
We don't forecast if you're infected, how seriously ill might you get.
We're just letting you know what's happening around you.
and your symptoms reports wouldn't contribute information in that regard,
but there are other ways that you contribute.
This is overusing the way's analogy.
But, for example, the fact that you were able to transit from point A to point B
down a segment of highway with nothing happening to you still tells us something.
One trivial example of that.
Let's say Sonal is your friend and you are my friend.
And Sonal gets respiratory disease and she has symptoms.
She gives it to you and you do not have symptoms.
but you give it to me. Through you, even though you have no symptoms,
the app will be able to take advantage of the information that Sonal has symptoms.
So by connecting us through a geodesic of tooth, you know, she is my friend's friend.
Even though you don't have any symptoms, your presence in the system is contributing information
or can help me be situationally aware of what's happening around me.
Because if the network gets dense enough, even if you have a high rate of asymptomatic,
you could actually impute infected non-symptomatic individuals.
Yes, that's exactly right.
Stepping back, the distinction between advanced warning and rapid detection is important.
There is a difference between early warning and rapid detection.
Most technology nowadays that tries to monitor the course of an epidemic,
it's like the CDC sits in Atlanta.
It has a distributed arrangement with testing facilities around the country.
Patients go and get tested for influenza-like illness, ILI.
Those test reports are sent back to Atlanta, and then two or three or four weeks from now, the CDC knows where the epidemic is today.
With the way's analogy, these are driving down the highway, and there's no traffic jam where you are right now.
There's a traffic gem coming up in the future.
You can avoid that traffic jam.
And that goes back to the three different time frames for like the CDC know in three weeks what was happening today.
You can use big data techniques about what's happening today to forecast to know what's happening.
now. For example, the Google Flu Trends idea, which, you know, has come under criticism.
And also this paper we have that looks at mobility data. You can look at the flows of people out of
Wuhan and you see that, oh, if many people from Wuhan go to these prefecture versus few people
from Wuhan go to that prefecture, this first one is going to have many more cases versus
lastly, being able to exploit some understanding about spreading processes on graphs in order
to be able to forecast using this canaries in a coal mine idea, what's going to happen in the
future. And the deployment of resources is not a trivial thing. If you're a public health official
and you're trying to see, where do I send my PPE, like the ventilators example,
could have known two or three weeks early that, nope, California is going to be okay, but New York's
going to be hit. We can reposition our equipment, or if you're a manufacturer and you need to
forecast demand, then knowing who's going to want these products in Texas versus Illinois is
helpful. What we're doing is a kind of decentralized collective effort. Some people might confuse it
with a facility that is being offered by these big tech companies. It's like ways versus having
police satellites above the city, you know, tracking where all the cars versus each person
contributes some information voluntarily and anonymously. There are a bunch of apps out there now
already, which are sort of citizen science apps, like tell us what's happening with you so we can
monitor the public. And that's great. And people can do that if they want.
But with us, they get something of advantage to themselves, mainly an assessment of their own risk.
It's not just a pro bono publico.
It's like the difference between dial the number and tell us where the pothole is so that the
municipality can go out and fill the pothole versus inform other people on ways that there's a pothole
so we all benefit from knowing where the pothole is.
I'm very tech friendly, but I've been very worried at the use of these technologies in the
hands of authoritarian governments.
And in some ways, I am more comfortable having Google and Apple know about me as a private
company than I would be about the United States government knowing. So I think we have to keep
our eyes on the ball is how can we use these technologies? Yes, in a way that's helpful to public
health, yes, but without not only not violating our civil liberties today, but without empowering
the abandonment of civil liberties in the future. On the work that you've done with the Wuhan data,
given that last point you made, is something akin to that even possible in the United States?
Yeah, so the Wuhan data project, that was a project where we did it in partnership with some Chinese colleagues and colleagues in Hong Kong, and they had a partnership with one of the primary telecos in China, and we got data on population movements out of Wuhan. So between January 1st and January 24th, we had data on 12 million transits through Wuhan, spreading out to 296 other prefectures throughout the rest of China. So we know that 100,000 people went from Wuhan to this.
prefecture and 10,000 from Wuhan to that prefecture and none from Wuhan to this other
prefecture. And then we had follow-up data from the Chinese CDC on COVID cases through February
the 19th. And basically, we were able to perfectly predict the location, timing, and intensity
of the coronavirus epidemic in each of the remaining 296 prefectures of China based on
population outflow. Incidentally, that allowed us also to validate the accuracy of the Chinese data
subject to a couple of provisos.
For example, we cannot validate the size of the number of cases in Wuhan,
which is probably much higher than the reported number,
partly because at the beginning they couldn't even know who was sick
and maybe they fudge the data a little bit too in Wuhan.
But certainly for the remaining prefectures,
it would have been very difficult to create the pattern in the data that we saw deliberately.
So yes, only some nations would be able to commandeer
such data from the private sector and then use it. But it doesn't need to be phone data. It could
be car data, for example, or train ticketing data, or car tolling data, or all kinds of other data
that tracks movement. And once again, it could be voluntary. People could report they're taking
a trip from point A to point B. When enough people start reporting, they're taking a trip from point A to
Point B, you can actually use that collectively generated information outside the government
to make predictions about where the intensity of the epidemic is likely to be.
So, yes, you're right that you do need some source of data.
If you're going to take advantage of these big data techniques, it doesn't have to be phone data.
It doesn't have to be in the hands of the government.
But, of course, you do need data to actually make it work.
In the context of the work you've done, whether it's with Wuhan, with the stuff you're doing with the Hunala app,
where does what I would call ground truth fit into this?
In other words, this idea that it is a knowable fact
as to whether or not someone has a virus.
That's an objective truth.
It's, you know, in some level...
It's an objective truth, but even that,
if you use the Heisenberg Uncertaining Principle,
you know, not all tests are perfect.
Fair.
There is an objective reality out there.
That's 100% true.
I agree with that.
Exactly.
So there's an objective reality out there
as to where their virus is present
or not present in an individual.
and the best approximation we have to that objective reality is a test, right?
And the tests have all their imperfections, including not on how the test performs,
but whether or not they are in fact available.
Where does testing fit on the one side?
And on the other side, where does testing fit at the aggregate population level
with the kinds of things you did with Wuhan?
Okay, well, the first question, we do allow people to tell us if they've had a COVID test
and if the test was positive or negative.
But stepping back, even the CDC's tracking of so-called,
called ILA, influenza-like illness, or other outbreaks, for example, when the WHO tracks
outbreaks around the world, there's a rush to define what counts as a case, even in the absence
of a test. So I can say, if you have a fever of above on a 1.4, and you have a cough, and you
were in China, then I'm going to say that you had COVID, even if we don't have a test on you,
Right. Or for example, if we're trying to get people to self-report whether they had the flu, if I asked you, you had something last week, was it a cold or was it the flu? There are non-test-based, non-biomedical test-based questions. You can ask people to tell the difference between conditions. You know, in medicine, we call this taking a history. And in fact, there's a joke when I was in medical school that we were taught, you know, that 80% of the diagnostic information comes from just talking to the patient. And 15% of,
comes from examining the patient, and only 5% comes from testing the patient. By the time you get to
the test stage, typically you're just confirming what you think the patient has. Even in your example
of the respiratory syncytrial virus, a case that we were talking about earlier, even in that case,
many doctors will know that this patient likely has RSV and will do the test to confirm their
suspicion. And sometimes a test will refute what the doctor thought. But nevertheless, a lot of
the variance is going to be explained just by the history and the physical. So it's possible to
certainly at the aggregate level, but even at the individual level, to use patient reports to get
some sense of what's happening. Now, we won't necessarily be able to tell the difference between
you had the influenza virus caused your respiratory problems last week versus a mild case of coronavirus
caused it last week. But in some sense, we don't really need to know that. From a public health
point of view, I would argue, that is not as crucial. It is an interesting point, that last point you make,
which is at sub-level, the question you're trying to answer isn't COVID yes, no, is,
am I suffering from a respiratory disease? Yes, no. It's a little bit like the shift that's
taking place right now around the country on just quantifying excess deaths from any cause.
Yes. Because there are a lot of people that die of coronavirus, and we never diagnosed them,
or they had a heart attack, or we thought they died of the flu, but actually it was coronavirus.
And you can look at these excess deaths and look at the impact of the epidemic in our population in some sense.
So I would argue that that's another metric.
And incidentally, just speaking for myself,
the only numbers that I've been following the last two or three months are deaths.
I've not been following a number of tests.
I've not been following percent positive.
I've not been following the case count.
And even deaths can be either an undercount or an overcount for various reasons.
But I regard that as a harder endpoint, even if it's a lagging indicator of where we are.
Given that you're looking at deaths, and obviously you're spending a lot of time thinking about this problem and looking at data,
what's your personal take on how bad is this bug?
You know, what is the CFR, what is the IFR?
Have you thought through that?
Yes.
I do have an opinion.
I think that case fatality rate is going to be between 0.3 and 1%.
So it'll be at least three times as deadly as the flu on average
and maybe 10 times as deadly.
There's a lot we can discuss about that.
Like, of course, the older the population, the higher the case fatality rate.
There's some intriguing evidence about
different viral strains being more or less deadly.
Possibility I had excluded even two months ago,
I didn't think that was very likely.
But there's some worrisome evidence that there may be different strains,
some of which are much worse.
And this is one theory that may partially explain why New York was harder hit than California.
I do not think by any means that's the only reason,
but it may partially contribute because that New York was seated by the Italian strain
and California and Seattle by the Chinese strain.
But the CFR is the probability.
of death, given that you came to medical attention.
And that's, some people don't seek medical care.
Some people can't get medical care.
So that definition's a little fuzzy.
Then there's something called the SCFR, the symptomatic case fatality rate.
That's the probability you die given that you ever have symptoms, which is a bit better
in the sense that you might have symptoms and never seek medical care, but we're going
to count you then.
Conditional on symptoms, do you die?
And then even better, from my perspective, if it can be known, if we have testing, for
example, is the IFR, which is the infection fatality rate, the fraction of people who are infected
that ultimately die. Now, here, the numbers range from 40 to 60 percent, but it's about half in my
view. The people who get this disease, for very weird reasons, never get symptoms. That means the
IFR is about half the CFR. So I think the IFR is going to be between 0.15 and 0.5, let's say,
which is a bad disease, by the way, just to be very clear. These numbers seem small, but
I think it's a serious pathogen, regardless of what fraction of people it kills.
Where do you think from your desk?
Yeah, I would say, I think this is likely to become endemic.
So that means, you know, this becomes effectively seasonal.
And, you know, I think a big important question is, how do we get from here to there?
And there being, you know, sort of the next rebound when it comes.
And any effort to open up the economy will be met with resurgence of disease and
emergence of new hotspots. I don't think there will be a silver bullet. But from a surveillance
standpoint, it looks like we have multiple weapons that we can deploy. We're taken together. We will
hopefully get some level of ability to fly with instrumentation to use your analogy. Without that,
we will continue to fly blind. One of the important aspects is whether or not we can effectively
run a process for testing, tracing, tracking, and ultimately isolating individuals that are
affected. We are way under where we need to be. In fact, there's an op-ed that was recently written
by a group of scientists, Sri Kusuri at Octan Bio, Jason Kelly at Ginko Biow Works, Fong-Zang
at the Broad Institute, and Jay Shender at the University of Washington. These are leading
scientists and genomicists and biologists who are essentially making the argument, it's almost a call
to arms to say, look, we have a lot of the technology in place. We just have to put the pieces
together so that we can do this at the scale at which it's necessary. It's almost like a call for
mobilization. So on the testing side, there are several efforts to do that. No test is perfect, but that
gives you the ability to say, okay, this infected individual could either seek treatment or should be
isolated. But as we know, before that individual knew that they were affected, they could have
infected other individuals. And this is where the tracing and tracking aspect is very important.
It's this idea of, if I know someone is now infected, can I essentially go back in time
and figure out who else they may have infected?
The tracking and tracing side, people essentially contribute their information in terms
of being symptomatic.
There's a Kinsa, for example, that has an internet-connected thermometer where they can
create a heat map so you can sort of track at some level the presence or absence of disease
or at least a proxy for the presence or absence of disease
by looking at above or below normal levels of temperatures.
So there are several efforts to do this.
The folks at Google and Apple have developed technology
where potentially you could use Bluetooth to warn you,
if you've been around someone that's been determined to be infected,
but these are all very backwards looking.
So it's almost like watching the weather forecast
a week after you've experienced that weather.
And so really what we need is largely through this testing,
tracking and tracing approach. And then you use the information in terms of who they have been
in contact with to figure out who are the most likely nodes in their immediate network so that
we can get those individuals tested. And once you do that in aggregate, that's what gives you
sort of the living weather map. That's very actionable from a policy standpoint, whether it's,
you know, at the governmental level, but also figuring out, am I employees safe? Can I open my
business. Can I go back to work? Can I expose my children to go back to school? Yeah, I mean,
I agree 100%. I think I'd add two things to what you're saying. There's no way that we can resume
a normal economy unless we unfortunately are willing to tolerate some risk. And for example,
this applies most directly to the reopening of schools and businesses. So I think we're going
to need some kind of universal liability waiver, some kind of change in laws, ideally coupled with
some kind of no-fault insurance so that, you know, we can compensate people who fall ill or
die. Second, my lab does a lot of work on social contagion, and we've been experimentally working,
we've been doing field experiments with support from the Gates Foundation, with support from
the NOMIS Foundation in Switzerland, the Robert Wood Johnson Foundation, with the Tata conglomerate.
We've been doing experiments in India and in Honduras, where we try to create artificial tipping
points in social behavior. So how do we change vaccine behavior? How do we change breastfeeding
behavior? How do we change latrine behavior? Public health behaviors in these settings? Well,
the same thing is relevant, of course, in our country. How are we going to change mask wearing
behavior? You know, how do we get people to keep six feet distance, for example? How do you get people
to wash their hands? There are all these actions that individuals can take that reduce collective
risk. And so one of the things we've been doing is applying some of our ideas and techniques and
mathematical tricks that we have developed
for changing population behavior
at scale in these developing
world settings to how can we apply them
in the United States? You talk about
social contagion. I'd be
curious to get your take on two fronts. Number
one, obviously, when
it comes to something like a coronavirus,
the contagion here is
physical, right? You have, you know,
it's the proximity to the individual where the
virus can jump from person to person.
I would imagine that
sort of a central node in the network
infects more and more nodes because they have higher connectivity.
How does the concept account for the possibility,
if such a thing in fact exists, of a super spreader?
So we know they're super spreading events with all epidemics.
Typhoid Mary, for instance, was a super spreader.
The question is what explains it?
A couple of things when thinking about super spreaders
is important to distinguish.
First is in any distribution, they're going to be outliers.
So, you know, we have a thousand people.
So let's say the effective reproductive rate of the pathogen is, let's say, two.
And it either could be that every single person has the same reproductive rate,
everyone in expectation infects two people, or we could have some range.
Some people infect nobody.
They're zero.
Some people infect one person, some two.
And one or two guys infect 100 or 300 people.
They could be super spreaders, but maybe they're just super spreaders by chance.
There's nothing about them or that they did or their environment that made them a super spreader.
there's just a natural distribution, and someone has to be at the top and someone has to be at the bottom.
But sometimes there's a more interesting biological or social explanation for super spreading.
So, for example, in many of the cases of coronavirus, we've seen super spreading events
that have typically occurred in very compact situations, where a sick person was in a contained
environment with many other people. If that same person had gone to some other situation,
they might not have caused a big super spreading event, but it was something about the environment.
maybe the air conditioning or the density of people in the room
or certain other kinds of things that might prompt that individual
in that setting to be a super spreader.
Or it could be something about the person or their behavior.
For example, you might have more sensitive nerves in your throat
and your nose so that when you get infected with a pathogen,
you cough and sneeze a lot.
I get the same pathogen.
I'm otherwise identical to you, but I don't cough and sneeze a lot.
So you spread to many people and I don't.
And it's just something about you, your body,
that is causing this event.
And sometimes it's even the pathogen,
like the strain of the virus you get,
might actually be more transmissible
than the strain I got, for example.
So it's the pathogen, the host, the environment, or chance
are the four things that might possibly explain
super spreading events.
Can you detect super spreaders in a network?
And I think the answer is probably yes.
Yes.
Can you model that and forecast that
in any sort of predictable way?
I'm reasonably sure that you can forecast
the role of super spreaders and the existence of super spreaders in an observed epidemic.
What I don't know is if you can forecast who will be a super spreader.
I suppose there are, as we've been discussing, certain things about the individual that you could maybe tell
or certain things about the environment that you could tell.
You can know basic things like, for example, people have many connections.
I don't think it's a coincidence that early on all these very prominent people, actors and politicians,
got the pathogen, people said, oh, it's because they're in the public eye or they're rich and they
had access to tests and that's how we knew it. Yes, those things are true, but also those people
are shaking a lot of hands. They have high degree. They're connected to many people. And they also are
acting as a forecast of who's going to get the disease. So, Nicholas, if we're relying on users to
essentially self-report how they're feeling, essentially self-report who their contacts are, there's a
level of subjectivity there, isn't there a risk that at some level this approach would be
vulnerable to a bad actor that could somehow meaningfully skew the data? Yeah, so there's always
that risk of someone, there's always the risk of bad actors. You know, there was that very funny
artist, I think he was in Germany or somewhere in Europe where he got like a hundred cell phones
and put them on a little wagon. Yeah, and create traffic jams. Yes, and drag them through the
city and created traffic jams on Google Maps. And so, you know, I suppose you could have people
like that. And of course, there were always people like typhoid Mary, you know, was actually a cook
and refused quarantine and kept infecting people until she was finally the state basically
incarcerated her. So yes, there could be people who act in a way. But they are, as I discuss
actually in my book blueprint, I'm quite an optimist about our species. We are a species that's
capable of love and friendship and cooperation and teaching. And of course,
we have evil and violence and selfishness and hatred and we have had those since time and
memorial, but I believe the good qualities outweigh the bad qualities. In fact, I show that
provide a lot of compelling evidence in Blueprint. And so yes, you're right that there could be
bad actors, but I think they will be swamped by the good actors. But I do want to go back to
this business about subjective versus objective because I see a problem where both the far left
in the far right, for their own reasons, wish to deny the existence of an objective reality.
And it's enraging to a scientist like me. So, you know, you see people who will say, oh, well,
it's just going to disappear, you know, or it's nothing, or why do we need to listen? But denial
of the existence of an objective reality doesn't make that objective reality go away. It's still
there. So, Nicholas, thank you for joining the A16D podcast. And I, like you, share incredible hope
that while we will be living with this virus for a long time, we will find a way to live with
this virus going forward. No, you're welcome. I mean, I've had a really wide-ranging and
interesting conversation. Thank you for having me.