Life Kit - Fake News Can Be Deadly. Here's How To Spot It.
Episode Date: April 20, 2020Experts say the coronavirus outbreak may be the biggest source of Internet misinformation ever. Fake cures, unscientific tips, and conspiracy theories about COVID-19 are flooding the Internet — but ...there are ways to sniff out misinformation.Learn more about sponsor message choices: podcastchoices.com/adchoicesNPR Privacy Policy
Transcript
Discussion (0)
Lately on the NPR Politics Podcast, we're talking about a big question.
How much can one guy change?
They want change.
What will change look like?
More energy.
Drill, baby, drill.
Schools.
Take the Department of Education closer.
Health care.
Better and less expensive.
Follow coverage of a changing country.
Promises made, promises kept.
We're going to keep our promises.
On the NPR Politics Podcast.
Hi, LifeKit. My name is Laurel Scarnelli.
So I have a tip, and it doesn't necessarily correlate to the coronavirus or quarantine,
but hey, it may come in handy if this thing lasts a little bit longer.
Every year I keep part of my garden free for the weeds to grow because I find them so beautiful.
And, you know, the colors are great and the textures are great. And instead of going to the grocery store to buy flowers in
the summer, I kind of just pick the weeds in my garden. Okay, thank you. Bye.
This is your NPR Life Kit on information, and maybe more importantly,
false information, and how to see through it and avoid it.
This is something that's obviously taking on new importance
during the coronavirus health crisis.
I'm Myles Parks with the NPR Politics team.
Health officials right now aren't just having to battle an illness
with no known cure or vaccine.
They're having to fight back against internet trolls
and conspiracy theorists.
The World Health Organization has labeled it an infodemic, as in information-demic.
Here's the WHO director general.
While the virus spreads, misinformation makes the job of our heroic health workers even harder.
It's diverting the attention of decision makers,
and it causes confusion and spreads fear to the general public.
Experts say this outbreak may be the biggest source of internet misinformation ever. The virus
was not created in a lab as a bioweapon, and inhaling hot air from a hairdryer is not a cure.
You've probably come across a piece of bad information about the virus. You might not
even know it.
We want to help.
We put together this guide to sniffing out misinformation before the current crisis.
But all the tips hold true, and they're even more important now.
The guide starts with a story that shows just how high the stakes are when bad information spreads.
Caitlin Dickerson is an immigration reporter for the New York Times.
She covered this story. It took place in Twin Falls, Idaho, before the term fake news was even in the national vocabulary.
City council meetings are usually pretty low stakes.
But at this meeting in the summer of 2016 period, where anybody can come up to the microphone
and just offer comments on what's happening around
town. A series of people get up one after the other. All asking about the same thing. They say
a horrible crime has been committed in their community. And they start to ask about a sexual
assault case involving refugees, Muslim refugees. It's an appalling story. Basically the people in
Twin Falls are concerned about a
five-year-old who people say was assaulted by a group of Syrian refugees. But at the meeting,
the city council members have no idea what they're talking about. There's lots of confusion.
The local government officials say they'll look into it. But over the next few days and weeks,
it goes from these sort of fringe, highly partisan blogs into the national media.
It goes viral at a time of intense fears about refugee resettlement and Islamic terrorism.
People from all over the country and even the world are accusing the government of a cover-up
operation. Members of the local government, the mayor, the city council members, local judges, the county prosecutor.
They were basically inundated for months on end with threats, violent threats,
very visceral and descriptive threats from all over the world.
The mayor's wife received a disturbing threat on her work voicemail.
It was in a monotone, sort of slow and deliberate voice.
He basically asked the mayor's wife, you know,
what would it be like for you if you were raped and assaulted and nobody paid any attention to you?
Very, very scary and unsettling stuff.
All of this outrage, it was based around something that didn't even happen.
It's a false set of facts.
There were no Syrian refugees resettled in Twin Falls at the
time. The alleged assault that took place, it involved three kids, a five-year-old, a seven-year-old,
and a 10-year-old videotaping on another kid's phone. Disturbing? Yes. But nothing like the
sort of violence that was circulating on the internet and in the minds of the people threatening
the city officials. The story tapped into people's worst fears, of the people threatening the city officials.
The story tapped into people's worst fears, and it took over the city for months.
Bad information, whether it's lies, propaganda, or just ignorance, it can have real-world impacts.
If you want to be a member of a functioning democracy, you've got to avoid it.
And we want to help you. Avoid taking it in and avoid sharing it.
For every headline, there's also another story about the people living those headlines.
On weekdays, Up First brings you the day's biggest news. On Sundays, we bring you closer with a single story about the people, places, and moments reshaping our world.
Your news made personal.
Every Sunday on the Up First podcast from NPR.
In college, Mustafa Suleiman started a helpline for young British Muslims.
People were just looking to find support in a language that made sense to them.
Today, he's CEO of Microsoft AI, where he's building digital helpers.
Think of me as your superpower in your pocket.
Building the future of AI.
That's on the TED Radio Hour podcast from NPR.
What's in store for the music, TV and film industries for 2025? We don't know,
but we're making some fun, bold predictions for the new year.
Listen now to the pop culture happy hour podcast from NPR.
So as an immigration reporter, Caitlin runs into misinformation all the time. We'll get into why
that is in just a bit. But before we get
into the nitty gritty, I wanted her to give her biggest overarching tip for avoiding wrong
information. The thing you should bring with you every time you go online and every time you pick
up something to read. I think the most important thing when you pick up a news story is to read it
with skepticism. And I don't view that as a judgment of the reporter who wrote
the story or of the outlet that published the story. I just think that's a smart way to read
the news. It's the way that I've always done it. It's the way that I hope people do it,
including when they're reading my stories. There's a fine line between that healthy skepticism
and making sure it doesn't turn into cynicism about the truth. Caitlin isn't saying you can't
trust anything. She's just saying don't follow anything blindly and unquestioningly.
A really good sign for me, one that makes me feel comfortable, you know,
with what I'm reading and believing what I'm reading is when a reporter shows their work.
Not only do they tell me what it is they know, but they tell me how they know it.
They tell me who told them.
A lot of important stories use anonymous sources.
That doesn't have to be a red flag all the time.
People who are close to a story might just not be authorized to speak about it publicly.
That doesn't mean what they're saying isn't true.
I at least want to know what that person's job is, what's their connection to the story itself, and how do they know the information.
So let's start there with your first tip.
Wherever you're getting your information, be skeptical.
Ask the source to prove their work. And in general, it's probably good to take in information from a bunch of
different sources, even if you have a favorite. So I want to take a step back at this point and
look at the bigger picture. Your second tip is a little bit of a bigger ask. Understand the
misinformation landscape.
For help on this, I went to Carl Bergstrom and Jevin West.
They teach an overwhelmingly popular class at the University of Washington.
The topic is relevant, but it's actually probably the title that attracts the swaths of 18 to 22-year-olds. Are we allowed to use the term?
For our public radio purposes, we'll call it calling BS.
Let's go BS to be safe.
Yeah, totally.
There's something inside me dying.
It's okay.
Yeah, let's do BS to be safe.
They started planning the class long before the 2016 election that put information consumption front and center in America.
After all, the idea of BS or bad information, it isn't
new, explains Carl. I'm an evolutionary biologist, and this notion of misinformation predates humans.
Take the raven, for example. When a raven has food, it likes to hide some of it for later,
also known as caching. But it'll look around first and see if anyone's watching it. And if
other ravens are watching it, then it'll do what's called fake caching.
The bird will actually fake like it's burying its food and then just move along.
So that when they try to steal it, it's gone off and hidden it somewhere else.
So we humans aren't the only ones trying to deceive each other.
But Jevin also says when it comes to bad information, there's a key difference between people who are outright lying versus people who are peddling BS.
Liars actually know the truth and they're just sort of moving you away from the truth.
Whereas people that BS, you know, don't really care so much about the truth as much as wanting to impress and persuade you and to sort of grab your attention.
Which is where in the modern age, the internet
comes in. The 2016 presidential election was followed by this huge reckoning. The U.S.
intelligence agencies agreed that Russia attempted to influence the results, which also isn't new for
Russia. The country has engaged in informational warfare, also known sometimes as active measures,
for decades. What was new was that the American public was this giant target
all tied to a platform that was tailor-made to spread misinformation like wildfire.
Social media.
Carl says the reason social media is such an effective tool for spreading BS
is that in some way, it's in the best interest of the Facebooks,
the Twitters, and the YouTubes of the world
to allow some level of this persuasive content, regardless of the truth. They're for-profit
companies, after all. They don't care about what you're seeing. They don't care if you're watching
a video that explains how the earth is flat. They only care that you're on the platform. They just
want your attention. And I think they're sort of like the grandest of BSers. Carl says the sites are basically conducting science experiments on our brains, figuring out what will be the most addictive.
The content that we're delivered has been curated and selected by a set of machine learning algorithms that are basically running large scale experiments on all of the users of the platform to see what keeps people clicking, what keeps people on the site. The big social media sites have improved their platform since 2016,
deleting hundreds of millions of fake accounts, adding fact-checking widgets.
But the fundamental problems remain. Their business model
depends on engagement, not on any sort of obligation to the truth.
I'm wondering why the answer isn't just use social media less.
Well, for sure. I mean, that's something I advocate in my personal life and also even in
the classroom. I wouldn't want to go back to a world without it. I think it provides a lot of
good things for the world, but I think we are using it too much. So your second tip,
understand the environment. Misinformation isn't new, but the platforms for engaging with it are.
We'll get into some more tips about how you can use social media in smart ways,
but maybe the first thing to think about is just maybe use it less.
Okay, so misinformation isn't new. That also means it can be sort of predictable. Our third
tip is to recognize the types of topics that draw a lot of false
information and be vigilant when you're consuming news or media about those topics.
So I want to come back to something we touched on earlier. Caitlin covers immigration for the
New York Times. She says she runs into misinformation all the time, whether it's
online or from politicians or from advocacy groups. It's frustrating.
You know, it's something that both Democratic and Republican politicians do.
And it's something that totally predates the Trump administration, although it's probably
reaching a peak.
I asked her exactly why it is that immigration is such a magnet for bad info.
It's clear she's thought about that question a lot.
And she breaks it down into two key factors.
One is that immigration is incredibly complicated.
You know, lawyers who study it compare it to the tax code.
People don't have time. You know, people have lives. They have jobs.
They don't have time to familiarize themselves with the nuances of immigration law.
And I don't begrudge that, but I think it becomes easy then for activists to capitalize on the lack of information
and boil it down in ways that aren't fully accurate.
When she mentions activists, she's talking about anyone who has a stake on any side of the issue.
These are the people who want to persuade, and that means they can have a lot of motivation to want to bend the truth.
This is broader than immigration, too.
It's true for any topic with complicated policies.
Voting, for instance, is something that's a little different in every single state,
so it leaves open the door for a lot of misinformation.
The other thing is that immigration is just this incredibly emotional issue.
It relates to race. It relates to demographics. It relates to religion.
It relates to people's feelings about the national identity as well as their personal identity.
And so it's this emotional hot-button issue that people get really excited about on every side of the political spectrum.
And that emotional response should be a red flag.
That's something I talked about with Peter Adams.
He's the senior vice president of the News Literacy Project.
He says that if you're having that emotional response,
it actually means you should double check whatever information you're getting,
whether it's from a politician's stump speech or from reading your news feed.
If it makes your blood boil, check it. If something is causing you to be fearful or outraged and to experience strong emotions like that,
take a moment and maybe do a quick web search to see if it's been debunked
or if anyone else is reporting it who is credible. The same goes for stories that are rapidly
changing or developing. Peters says people who make money from clickbait or fake news stories
don't need to follow the same ethical guidelines that media organizations do.
Propagandists and disinformation agents and trolls and just chaos mongers seek to sort of shoot the curiosity gap or get something out in the first 15 minutes before legitimate news outlets are even able to get information and verify it and get it out.
News outlets aspire to standards, right?
And they have to verify information before they share it.
And that takes time.
Sometimes it takes five or ten minutes, which is pretty quick,
but in that five or ten minutes, other folks with bad intentions online can also push a lot of garbage.
So tip number three, recognize the times when bad info works best.
Just ask yourself, is this a complicated subject?
Is it something that's hitting on my emotional triggers?
Or is this a breaking news story that's going to change?
Okay, so when you see a story about gun ownership or reproductive rights or immigration,
you're going to be on the lookout.
But what do you do when you sniff it out?
How are you going to double-check what you're reading?
Peter's group wanted to answer exactly that.
So they're working on an app called Informable.
If you're really well-read or well-listened in the case of podcasts, you might be thinking you can tune out at this point.
That the people who fall for this stuff are sort of simpletons. Peter says not so fast.
We all have our blind spots, especially if you consume a lot of information. It means you are
often sort of info grazing. So the bigger the news junkie, the more instances you might have
where you're quickly making a judgment, waiting in line at the grocery store, you know, refreshing
your timeline or whatever. And that actually puts you at a disadvantage if you're making decisions
about sharing and liking things. The app focuses on a few categories that you can use to model your
own thinking. I asked Peter to walk me through it. The first category is ad or not.
This is basically discerning whether something you're reading or seeing
is there because someone paid for you to see it.
If someone paid, then it's a biased source of information.
Telling the difference is especially tough on social media,
where organic posts, things from your family and friends,
get mixed with paid content from companies and even politicians.
We have, first off, a shot of a piece published on BuzzFeed.
And it says at the top, you know, in pretty prominent all caps, paid post.
And the headline is 10 ways to succeed in college, according to people who went to college.
And further down, it lists Course Hero as the, quote, brand publisher.
Second is evidence or not.
When a source makes a claim, you just need to question whether there's reliable information that proves their point.
If someone is talking about population, for instance, and they cite census data, that's evidence.
But as Peter's playing through this section in the app, he runs into a sketchy tweet. It has a picture of some Baby Ruth candy
bars and airheads and Cheez-Its and cans of tuna next to a cardboard box. And the claim is these
are the FEMA meals my brother received today. And meals is in quotes. And it has a hashtag,
you know, Hurricane Maria. This may well be true, but it doesn't really provide evidence that this
is in fact a FEMA meal. Anyone could take a picture of some baby Ruth and other candy and Cheez-Its and say this is a FEMA meal.
So that would be something that would warrant looking into further right before you sort of accept that as evidence.
Next, he runs into a very cute Instagram post that shows two dogs embracing.
And it says, you know, this dog hugs every other dog he sees during his walk.
It is just a still image of two dogs that appear to be in an embrace, but it is not evidence that
that dog in fact does that. Now this part was tough for me to swallow. I mean, I have to question
every time I see a cute puppy post, is nothing sacred anymore? I asked Peter about this after
he showed me a meme of some kids goofing off in the back of an auditorium.
The caption says, university students hard at work during a lecture.
And one of the kids is sort of half asleep on like three different chairs.
And another kid is watching Game of Thrones on a laptop.
You know, there's some people who would say it's entertainment at that point.
And if I'm laughing about what these two students are doing, supposedly in the back of this auditorium during class, does it matter to me whether they actually did that and it affirms a belief that they already
have, it strengthens a belief that they already have that, you know, kids today are lazy, that
they're not learning anything, that college is a waste of time, it may reinforce those beliefs.
So you, you know, for you, it might be funny because you don't necessarily have that belief,
but someone else who does, it may act as evidence for that.
Yeah. Like if I was, if I was was against public funding for universities or something like that,
then it would have this completely different effect on me than the person who originally shared it.
Exactly.
When it comes to evidence, numbers come up a lot.
The problem is that all numbers aren't created equal.
Here's Jevin from the University of Washington.
You can always remove one variable.
You can always only include men instead of women. There's all sorts of ways to cut up
numbers and tell stories. And that's the whole point that we want the public to be aware of,
that when they see a number to question, you know, where that number comes from.
When Caitlin is covering immigration, for instance, a number she has to deal with a lot
is the number of border crossings every month at the
U.S.-Mexico border. For example, there were months over the summer when tens of thousands of people
were crossing the border in a single month, you know, 60, 70, 80,000 people, which sounds like a
huge, huge amount. Those numbers were accurate, according to the Department of Homeland Security.
But how someone chooses to present them can vary really widely
depending on the story they're trying to tell.
There weren't enough resources along the border to actually be able to house
and care for people who were crossing in a safe way.
And we saw the effects of that.
But, you know, we also have had our reporters who cover the economy
do stories about how when you look at the overall population
and you look at the overall population and you look
at the economy it can in fact sustain those very large numbers of people and in many economic ways
we actually need them to support the economy so when you look at a number like 70 80 000 people
on its face and you think my god that's that's huge that's untenable we have to change something
and then you actually look at the numbers and find out,
well, actually, it's quite easy for those people to get jobs and to integrate into society.
That's another way of looking at it.
When it comes to numbers, context is the key.
A number like a city's murder rate, it might look big and scary,
but if you check what it was 10 years ago and you see that it's gone down substantially,
then your takeaway will be really different.
Or if you compare it to a neighboring community. Jevin and Carl teach their students this one quick trick, that when you're looking at a graph, check the x-axis and the y-axis
to make sure they're not overly zoomed in or zoomed out to show a convenient amount of data.
The important thing to recognize is that the person presenting the numbers does have,
even if they're true numbers,
does have a lot of power over controlling how you feel about them, how you respond about them. The last really good indicator for when something is suspect is whether you know the source. On
social media, that means knowing the creator of the meme, not just the person sharing it.
Let's say your Uncle Bert retweets a meme on Twitter. You see the picture and a notification indicating Uncle Bert re-shared it,
which your brain initially interprets as Uncle Bert being the source.
But you need to rewire to then either not take the meme as fact
or to follow the rabbit hole down to the creator.
If you can't get a firm grasp on who that is,
then Carl says you shouldn't recirculate it.
Uncle Bert was the one that retweeted it. But if then if you don't know the person that sent it,
then, you know, you don't really know where that information is coming from.
You know, so sharing information that you don't know where it's coming from is kind of like
picking up candy on the street and just eating it.
So that's your fourth tip. Ask some questions of what you're seeing and reading.
Is it paid for by a company or a politician or another biased source? Is there good evidence? And are the numbers in context? bad stuff sometimes. You see it on Facebook or at a family party, or you might hear about it at work.
It's just going to happen. I promise. We want to help you with that because it's important to value
the truth. But it's also really hard to tell someone they're wrong if it's about something
they care about. I think the most important thing is to recognize that when people believe strongly
in something, even if it's not borne out by the facts, there's a reason. You know, they've read something or they've been exposed to something,
they've been told something that was compelling to them and that seemed legitimate. And so it's
worth acknowledging that, you know, and that their beliefs may also be informed by their experiences
and challenges that they've faced. And so I think, you know, meeting people where they are
is the first and most important step to having a productive discussion over disagreement.
Devin and Carl said they're careful as they teach their course to walk a fine line.
We start with a few rules for calling BS that we really want to instill in our students because we don't want to create a legion of jerks or the well-actually guy or whatever.
But truth is important for our democracy and for our health. So we have to be able to talk
to each other. The first ground rule, if you're going to try and correct someone who's wrong,
is to make sure you're right.
So you want to make sure that you've got your facts straight and you don't leap in
and call BS when you're wrong. That's a sure way to undermine
your credibility. So once you've got your case mapped out, think about opening with common ground
and a question. Carl used the example of vaccines as one that's really charged, but also really
important. There's no scientific link between vaccines and autism, and yet a chunk of the
population believes there is. That's led to
outbreaks in recent years of diseases like measles. So if you're in the daycare pickup
and you start talking to somebody who starts telling you everything they know about the
negative consequences of vaccines, Jevin says start with the common ground.
It's really stressful being a parent. There's a lot of decisions you have to make.
And so, you know, what have you been reading? What have you been looking at? And then that'll give you a chance, you know, you can hear from them. You've got this common ground, you guys both really care about your kids. And then you can kind of share what you've learned and say, well, you know, from what I've learned, measles is not a minor disease. It's really serious. And I care about my kids a lot. And so since there aren't, there isn't this evidence, you know, that vaccination is causing autism, I was careful to get them vaccinated.
Just finding that common ground, if you set it up as like, you know, you're a dumb hippie,
you're a big pharma tool, then you're not going to get anywhere, right? But if you're both just,
you know, you care about your kids, then you can kind of go from there.
Okay, so clearly, this is not an easy conversation to have. And honestly,
when misinformation is online, it can be even tougher to engage.
Think about a comment section and how often these things just turn into this bickering back and forth instead of a productive conversation.
One of the problems with the discussions that take place on Twitter and Facebook is that they're typically public discussions.
And so instead of you taking me aside and saying, hey, Carl, look, like, you know, I understand why you think that that's actually not true. And here's here's why you're sort of very publicly saying, Carl, what you said is BS. And here's why. And you shouldn't have said that. And you're calling me out in public. And so, you know, my sort of seems that the informal first rule of discussion on the Internet is to always double down on your own stupidity.
And I think some of that is a response to being called out in public for getting things wrong.
It seems like it ups the ante.
It's like me going up to you and kind of shoving you and expecting you to react well to what I'm saying.
Exactly. I think that's right.
So it's not that you shouldn't engage when you see something hateful or wrong.
But if you're tempted to write a comment, maybe just think about sending an email instead.
Or if it's someone who you see regularly, think about bringing it up in person.
Which brings us to your last tip.
If you know you're right and you want to help correct misinformation, be humble.
Don't assume bad intentions or stupidity.
Just meet the other person where they are and be curious.
Try to have the conversation in person,
or at least in a private online setting.
So that's it.
You're an information pro now.
You're basically an encyclopedia.
Let's go back over the key points when it comes to finding good info.
Number one, be a little skeptical of everything you read and see.
Even if it's your favorite reporter or your favorite magazine,
it's probably best to take in news from a few different places too.
Tip number two, know the landscape.
Social media sites don't have to tell the truth,
so don't treat them like they do.
Tip number three, look out for red flag topics.
Be really careful if it's a complicated subject or a breaking news story or an emotional trigger.
A lot of misinformation exploits our values.
It exploits our patriotism.
It exploits our religious faith.
It exploits our dedication to ideals like equality.
Tip number four, take in information with a few questions at the top of your mind.
Is it paid for by a company or a politician or another biased source? Tip number four, take in information with a few questions at the top of your mind.
Is it paid for by a company or a politician or another biased source?
Is there good evidence?
Are the numbers in context?
Tip number five is basically don't be a jerk.
If someone believes wrong information about something, there's a reason for that.
Be curious and helpful. Subscribe to our newsletter at npr.org slash life kit newsletter. And here, as always, is a completely random tip.
This time from listener Viva Dadwall.
What's my life hack?
Well, it includes talking to strangers, including finding all sorts of excuses to be able to do so.
So sometimes I take a telescope and I set it up on a corner and invite people to come look up at the sky.
And it more than often works.
If you've got a good tip, leave us a voicemail at 202-216-9823 or email us at lifekit at npr.org.
This episode was produced by Sylvie Douglas. Megan Cain is the managing producer. Beth Donovan is
the senior editor. This episode was edited by Brett Neely.
Our digital editor is Beck Harlan, and our editorial assistant is Claire Schneider.
I'm Miles Parks. Thanks for listening. This message comes from Grammarly.
89% of business leaders say AI is a top priority.
The right choice is crucial,
which is why teams at one-third of Fortune 500 companies use Grammarly. With top-tier security credentials and 15 years of experience in responsible AI, Grammarly isn't just another AI communication assistant.
It's how companies like yours increase productivity while keeping data protected and private.
See why 70,000 teams trust Grammarly at grammarly.com slash enterprise.
Since the beginning of women's sports, there's been a struggle to define who qualifies for the women's category.
Tested, from NPR's Embedded podcast and CBC, takes you inside that struggle.
Listen to Tested, the series that was named one of the 10 best podcasts of 2024 by Apple, Vulture, and The New York Times.
It's season 20 of NPR's Embedded Podcast.
Every weekday, Up First gives you the news you need to start your day.
On the Sunday story from Up First, we slow down.
We bring you the best reporting from NPR journalists around the world,
all in one major story, 30 minutes or less.
Join me every Sunday on the Up First podcast to sit down with
the biggest stories from NPR.