Hidden Brain - I'm Right, You're Wrong

Episode Date: December 26, 2017

There are some topics about which it seems no amount of data will change people's minds: things like climate change, or restrictions on gun ownership. Neuroscientist Tali Sharot says that's actually f...or good reason. As a general rule, she says, it's better to stick to your beliefs and disregard new information that contradicts them. But this also means it's very difficult to change false beliefs. In this favorite episode from earlier this year, we look at how we process information, and why it's so hard to change our views.

Transcript
Discussion (0)
Starting point is 00:00:00 Hey there Shankar here. As we reach the end of 2017, there are four words you almost never hear, whether that's from people in the news or from friends and family. You're right, I'm wrong. We don't hear those words and we don't say those words because we all have elaborate mental defenses to keep threatening ideas at bay. These psychological defenses can give us a short-term boost, even as they expose us, to long-term risks. We're taking a bit of time off this week, so we thought we'd share a conversation from March with the neuroscientist Talley Sherrod. We found this episode helped us understand why it's so hard to listen to ideas that challenge
Starting point is 00:00:39 our beliefs. It also gave us some techniques to help us see the world with fewer biases and preconceptions. As you gather with friends and family to celebrate the New Year, we hope you'll find it useful too. This is Hidden Brain, I'm Shankar Vedenthu. There are certain issues where we just don't see eye to eye with people on the other side of the political aisle. Often, it feels like we don't just disagree over policy, we can't even agree on the facts. I think people are mostly just setting their ways. I would say I'm right objectively, the facts prove that there is global warming. I don't know how you can dispute it.
Starting point is 00:01:22 But everybody thinks they're right. I think they're wrong. I think they're making a huge mistake because security has to be a very big issue. My guest today has spent years studying the way we process information and why we often reach biased conclusions. She says it's surprisingly difficult for us to change one another's minds
Starting point is 00:01:41 no matter how much data we present. But just a little bit of emotion, that can go a long way. Tali Shahraut is a cognitive neuroscientist at University College London. In her upcoming book, The Influential Mind, she explores how our minds work, why we ignore the facts, and how you can get people to actually listen to the truth. Tally, welcome to Hidden Brain.
Starting point is 00:02:11 Thank you, great to be here. Some months ago, you were listening to a Republican presidential debate, and Candida Donald Trump was asked a question about the safety of childhood vaccines. Here's what he said. You take this little beautiful baby and you pump. I mean, it looks just like it's meant for a horse, not for a child. And we've had so many instances, people that work from it,
Starting point is 00:02:33 just the other day. Two years old, two and a half years old, a child, a beautiful child, went to have the vaccine and came back and a week later got a tremendous fever, got very, very sick, now is autistic. Tali, you're a mom. You have two small children. At the time, Trump said this, one of your kids was two years old, the other was seven weeks old. Describe your emotional reaction to what he said. Yes, so when I was listening to Trump at that debate,
Starting point is 00:03:03 I was already quite concerned because I was a mother of a small child, so you're always concerned about the health and safety of your kids, especially when they're only a few weeks old. And so it really tapped into this fear that I had and the anxiety that I already had. And when he talked about this huge syringe, a horse-sized syringe that was going to go into the baby in my mind, I could imagine this syringe, a horse-sized syringe that was going to go into the baby. In my mind, I could imagine this syringe inserted into my small little child and all the bad things that could happen.
Starting point is 00:03:33 And this was a very irrational reaction on my end because I know that there's not an actual link between autism and vaccines. But it's not enough to have the data. Ben Carson, Dr. Ben Carson, was on the other end. Well, let me put it this way. There have been numerous studies, and they have not demonstrated that there's any correlation between vaccinations and autism.
Starting point is 00:03:59 But that wasn't enough, because the data is not enough, and even if the data is based on very good science, it has to be communicated in a way that will really tap into people's needs, their desires. If people are afraid, we should address that. I'm curious when you sort of contrasted, you know, the weight of the evidence on the one hand, and this very powerful image of the horse syringe and your seven-week-old baby on the other hand. How did you talk yourself into trusting the data over that emotional image? What really helped is that I understood what was happening to me,
Starting point is 00:04:40 you know, the first instinct was like a stress and anxiety, but because this is what I study, I knew what my reaction was. I knew where it was coming from. I knew how it was gonna affect me, and I think that awareness helped me to put it aside and say, okay, I know that I'm anxious for the wrong reasons, and this is the action that I should take.
Starting point is 00:05:05 It's a little bit when you're on a plane and there's turbulence and you get scared. But telling yourself, I know that turbulence is not actually anything that's dangerous. I know the statistics on safety and planes and so on. It helps. It helps people reduce their anxiety. The facts don't always relieve our anxieties though. helps people reduce their anxiety.
Starting point is 00:05:28 The facts don't always relieve our anxieties, though. Sometimes, they only harden our views. Sometime ago, Tally did a study where she presented information to people who believe that climate change is real and to people who are skeptics. She found, for both groups, people strengthened their pre-existing beliefs when new information confirmed what they thought, but both groups, people strengthen their pre-existing beliefs when new information confirmed what they thought, but both groups ignored information when it challenged their views.
Starting point is 00:05:51 I asked Ali about this. Our psychological biases are the same across individuals on average. We all have what's known in a confirmation bias. A confirmation bias is our tendency to take in any kind of data that confirms our prior convictions and to disregard data that does not conform to what we already believe. And when we see data that doesn't conform to what we believe, what we do is we try to distance ourselves from it. We say, well, that data is not credible, right? It's not good evidence for what it's saying.
Starting point is 00:06:26 So we're trying to reframe it, to discredit it. I want to sort of push back at you just on one point, which is sometimes it seems to me that it actually might be rational to reject data that comes in. If I've seen gravity work all my life, it makes objects fall toward the ground. And you take me into a room where an object seems to be levitating. It seems to me the appropriate reaction is not okay, everything I knew about gravity was wrong, but let me try and understand this anomaly. Let me try and understand why this object is not doing what it's supposed to do. It seems to me and I'm clearly betraying my own beliefs
Starting point is 00:07:05 here, given the overwhelming weight of evidence that climate change is real, showing me data that one winter is colder than normal shouldn't make me change my overall view. Should it? You're absolutely right. So the way that people tend to update their beliefs is that they use the new information in light of what they already believe because that on average is the rational way to go. Exactly as you said, when we encounter some kind of information that really contradicts what we believe very, very strongly, on average that information is wrong. Right, so I give an example in my book where if someone comes in and says, I just saw pink elephants flying in the sky and I have a very strong belief, obviously I just saw pink elephants flying in the sky,
Starting point is 00:07:45 and I have a very strong belief, obviously, that no pink elephants fly in the sky. I would then think that they're either delusional or they're lying, and there's good reason for me to believe that. So it's actually the correct approach to assess data in light of what you believe. There's four factors that determine whether we're going to change our beliefs. Our old belief, our confidence in that old belief, the new piece of data and our confidence in that piece of data. And the further away the piece of data is from what you already believe, the less likely
Starting point is 00:08:13 it is to change your belief. And on average, as you go about the world, that is not a bad approach. However, it also means that it's really hard to change false beliefs. So if someone holds a belief very strongly, but it is a false belief, it's very hard to change it with data. When we come back, we're going to talk about some of the ways we can reshape the beliefs of other people or our beliefs, and try and answer the question, if throwing good information at people doesn't drive misinformation out of circulation, how do you get people to buy the truth?
Starting point is 00:08:53 Stay with us. In often fields as though fear is used to motivate us to act. The vast majority of Americans today do not feel safe. On Sunday, Americans woke up to a nightmare that's become mind-numbingly familiar. This could be the great trojan horse of all time. Politicians use fear to get us to vote. TV programs use fear to get us to keep watching. Public health officials use fear to get us to quit smoking. I ask Tally whether fear might be an effective way to persuade people to change their minds and maybe even their behavior.
Starting point is 00:09:39 Fear works in two situations. It works when people are ready, stressed out, and it also works when what you're trying to do is get someone not to do something, an inaction. For example, if you try to get someone not to vaccinate their kids, fear may work. If there's an apple that looks bad, I don't eat it. Fear is actually not such a good motivator for inducing action, while hope is a better motivator on average for motivating action. You talk about one study in your book where a hospital managed to get its workers to practice hand hygiene, to get staff members to wash their hands regularly.
Starting point is 00:10:23 But it turned out the most effective thing was in frightening the staff about the risks of transmitting infections. It was something else. So in a hospital on the East Coast, a camera was installed to see how often medical staff actually sanitize their hand before and after entering a patient's room.
Starting point is 00:10:41 And the medical staff knew that the camera was installed and yet only one And the medical staff knew that the camera was installed, and yet only one in ten medical staff sat as their hands before and after entering a patient room. But then an intervention was introduced, an electronic board that was put above each door, and it gave the medical staff in real time positive feedback. It showed them the percentage of medical staff that washed their hands in the current shift and the weekly at rate as well. So, anytime a medical staff will wash their hands and numbers will immediately go up and there will be a positive feedback saying, you know, good job. And that affected the likelihood of people washing
Starting point is 00:11:23 their hands significantly. It went up from 10% to 90% and it stayed there. Instead of using the normal approach, instead of saying, you know, you have to wash their hands because otherwise you'll spread the disease, basically instead of warning them of all the bad things that can happen in the future, which actually results in inaction, they gave them positive feedback. One important idea that Tally has explored is something known as the equality heuristic. It's a mental shortcut and it's really quite simple. We tend to assign equal weight to everyone's opinion.
Starting point is 00:11:58 But sometimes when there's an expert in the room, this mental shortcut can lead us a stray. Different people have different expertise and it's better to put more weight on people who are more knowledgeable or have more expertise in the domain that we're making the decision in. And there's been a study showing that this equality heuristic is something that people do around the world. So it's not something that people do only in democratic governments, but studies have been conducted in other countries such as China and Iran,
Starting point is 00:12:31 and there too, people go according to the equality heuristic. If they need to make a decision, they will get the opinions of quite a few individuals and then tell you them up, and that's how they make the decision, instead of actually using the person in the room who is more knowledgeable and has more expertise. But how can you know who the expert is? When your question is about cancer, you can turn to an oncologist. But for other questions, many people might claim to be knowledgeable. It turns out there's a clever technique to separate the experts from the pretenders. It's called the surprisingly popular vote.
Starting point is 00:13:09 Let me show you how it works with an example. I have two questions for you. The first is, what's the capital of Brazil? Okay, here's the second question. What do you think most people will say is the capital of Brazil? Maybe you thought the capital was Rio de Janeiro, but maybe you knew the right answer, Brazil By looking at the results of those two quick polls, I can tell you with no prior information that the capital of Brazil is Brazil How would I know that? Because if I think Rio de Janeiroissonario is the capital of Brazil, I will also think that most people think that.
Starting point is 00:13:48 However, if I think Brazil is a capital of Brazil, I will still think the Rio is the answer that most people will give. So what this ends up being is that Brazil will be an answer that's more popular than people expect. I have to say it took me a minute to understand how this idea works. It's complicated, but it really is very clever. Here's another example. Let's look at a math problem.
Starting point is 00:14:13 In a lake, there's a patch of lily pads. Each day, the patch doubles in size. If it takes 48 days for the patch to cover the entire lake, how long would it take for the patch to cover the entire lake, how long would it take for the patch to cover half the lake? The correct answer is 47 days, because the lilies double in area every day. If they fill the lake on day 48, they will fill half the lake on day 47. But the intuitive answer is the 24th day, because 24 is half a 48.
Starting point is 00:14:45 If you ask people what the answer is, most people will say 24, a few people will say 47. When you ask people what they think others will say, nearly everyone will predict that others will say 24. Since more people say the answer is 47, then expect others to say the answer is 47. 47 turns out to be the surprisingly popular vote answer and the correct one. It's basically a clever way to figure out who has expert or insider information. Betting markets work on a similar idea. If there's someone in a group who has knowledge about something and you give them a financial
Starting point is 00:15:21 incentive to come forward with that information, you can bring insider information to the surface. So the surprising popular vote tends to predict the correct answer in many different domains. It can be factual, it can be math, it can be even estimating the price of art. So they did a little study where they asked MIT students to estimate the price of art. Now, MIT students, for the most case, they don't really have a lot of knowledge about art.
Starting point is 00:15:51 And yet, when they asked them to estimate the price of art, and also to say what other people will think about that, and they took the surprising popular vote, the surprising popular vote, the spot on. I wrapped up my conversation with Talley by exploring one last idea about how we might convince others to listen to the truth. It had to do with a study of Princeton students who got their brain scanned while they listened to a story of a young woman named Annabelle. Producer René Clarre read an excerpt from that story.
Starting point is 00:16:22 I know everyone has some crazy prom stories, but well, just wait. I was a freshman in high school in Miami, Florida, and I'm new to the freshman scene. I'm new to the high school scene, I should say, and it's almost December, so I've been in high school for about three months, and this boy Charles asks me out. He's British, he's a junior, and he's really cute. But sort of shy, but just, well, it doesn't matter. So I say yes, I'm excited." The story goes on. It includes, as Tally writes, love, rejection, blood, alcohol, and a couple
Starting point is 00:16:57 of policemen, all the requirements of a bestseller. I asked Tally to tell me what happened in the minds of the Princeton students as they listen to that story. So this is a study that was conducted by Professor Hassan's group at Princeton. And what they showed is that when one person listens to another person's story, the activity in the two brains synchronize. So if you're listening to me, then it is possible that the activity of your brain looks a lot like the activity in my brain. And at the beginning, the activity in your brain listening to my brain, it will be delayed in a little bit, right? So I'm thinking about
Starting point is 00:17:37 something, I'm saying the words, you're perceiving the words, and then the activity in your brain will follow the activity in mind. However, they found in that study that after a while, the brain of the listener started proceeding the brain of the storyteller because the brain of the listener was predicting what the storyteller will say. So, if you looked at the listener brain, you could predict what the activity in animals brain would look like. But the idea of synchronization seemed to be important for influence.
Starting point is 00:18:11 Why? In another study that the same group did, they found that when people were listening to very strong speeches, what they found was the brains of the different people listening to those speeches started synchronizing. So if we all listen, for example, to Kennedy's famous moon speech, her brains would likely look very much alike. Those who came before us made certain that this country rolled the first waves of the industrial revolution, the first waves of the industrial revolution, the first waves of modern invention,
Starting point is 00:18:46 and the first wave of nuclear power. And this generation does not intend to founder in the backwash of the coming age of space. We mean to be a part of it. We mean to lead it. And this is not only in regions that are important for language and hearing, it's also in regions that are important for emotion, in regions that are important for what's known as free-rear-of-mind, our ability to think about what other people are thinking, in regions that are important for associations. And you try to think, well, what's common to all these influential speeches that can
Starting point is 00:19:21 cause so many people's activity to synchronize, And one of the most important things is emotion. If the storyteller or the person giving the speech is able to elicit emotion in the other person, then he's actually having somewhat of a control on that person's state of mind. So think about it like this. If you're very sad and I'm telling you a joke, well, you're sad. So you're not going to perceive the joke as I perceive it when I'm happy.
Starting point is 00:19:53 But if I'm able to first make you happy and then tell you the joke, well, then you'll perceive it more from my point of view. So by eliciting emotion, what you're able to do is change the perception of everything that comes after. To perceive information as the person who's giving the speech wants you to perceive it. So you can see how this coupling, this idea that the audience is mind and the speaker's minor in some ways coupled together. You can see how this could potentially be used to spread good information. You know, you have a great teacher in high school and you're captivated by the teacher and you're being pulled along by the story the teacher is telling you, maybe about history or maybe about geography.
Starting point is 00:20:36 But you can also see equally how the same thing can work in the opposite direction that you could be listening to a demagogue or you could be listening to somebody who has a very sort of captivating rhetorical style and this person could also lead you astray in just the same way that the great teacher can lead you to knowledge and to positive things. Absolutely. whether we will be influenced by one person or ignore another person, are the same whether the person has good intentions or bad intentions, right? The factors that affect whether your influential can be, can you enlist it in motion in the other person? Can you tell a story? Are you taking into account the state of mind of the person that's in front of you? Are you giving them data that confirms to their preconceived notions? All those factors that make one speech more influential than the other,
Starting point is 00:21:32 or more likely to create an impact, can be used for good and can be used for bad. Tali Sharad, I want to thank you for joining me on Hidden Brain today. Thank you so much for having me. today. Thank you so much for having me. This week's episode was produced by Maggie Penman and Raina Cohen and edited by Tara Boyle, a special shout out to our former intern Chloe Connelly, who interviewed the Liberals and Conservatives we heard at the start of this episode. Our staff includes Jenny Schmidt, Parts Shah and Renee Clark. This week our unsung hero is Brian Malford of National Public Media. NPM is the group that sells our sponsorship messages. Brian is a great mix of
Starting point is 00:22:10 liberal and conservative. He's always liberal in his encouragement and conservative in his promises. As the saying goes, he under promises and over delivers. Thanks Brian. I'm Shankar Vidantum and this is NPR.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.