Hidden Brain - Mind Reading 2.0: The Double Standard
Episode Date: February 15, 2022It’s easy to spot bias in other people, especially those with whom we disagree.  But it’s not so easy to recognize our own biases.  In the latest in our "Mind Reading 2.0" series, we revisit a f...avorite conversation with psychologist Emily Pronin. We'll look at one of the most bewildering aspects of how we read minds — in this case, our own. If you like this show, please check out our new podcast, My Unsung Hero! And if you’d like to support our work, you can do so at support.hiddenbrain.org.Â
Transcript
Discussion (0)
This is Hidden Brain, I'm Shankar Vedanta.
If you were to paint a portrait of the inner workings of your mind, what would it look like?
Maybe you would use bright splashes of color to represent your most intense emotions,
shades of grey to reflect the complexity and nuance of your thoughts.
Swirls and spirals to express the moments when
you are lost in rumination.
No one else could paint this picture. You know yourself better than anyone else and you
are the one most attuned to your inner rhythms, your fears and insecurities, your hopes and
dreams.
Yet, even though we may feel we are acutely aware
of every corner of our mental landscapes,
it turns out many aspects of our minds are hidden from us.
It's almost like if you imagine a fork in the road
and it just goes two different ways,
there are just two different paths here.
There's the path that we use for self-judgment
and there's the path that we use for self-judgment, and there's the path that we use for judging others.
And in my view, the path that we use for judging others
is we look at their actions.
The path that we use for judging ourselves
is we look inwards.
This week on Hidden Brain,
we bring you the third installment
of our Mind Reading 2.0 series.
In previous episodes, we explored how we read other people's intentions
and the social illusions that pervade everyday interpersonal relationships.
Today, in a favorite episode from 2020,
we look at one of the most bewildering aspects of how we read minds.
In this case, our own.
On a daily basis, all of us evaluate others. We think about the claims of people who want
to sell us something, we gauge the ideas of colleagues, we assess friends and family.
We also regularly look into our own hearts and minds,
we evaluate ourselves.
At Princeton University, psychologist Emily Pronen has studied
why our minds come to very different conclusions about ourselves and others.
Emily Pronen, welcome to Hidden Brain.
Thank you, Sean Carr. So a few years ago, Emily, you conducted an experiment
where you brought volunteers into a lab and you told them about a range of different biases,
you know, biases like the halo effect where, you know, you see someone who's very beautiful
and you assume this person must also be very intelligent or a bias like confirmation bias where we go looking for information that supports our pre-existing
views. And you did something very interesting. You asked the volunteers whether they thought
that they would fall prey to these biases. What did they tell you?
We had students in a class so that they kind of all knew each other from being in the class
together. And what we did is we described each bias just in a few so that they kind of all knew each other from being in the class together.
And what we did is we described each bias just in a few sentences.
We didn't use the word bias. We didn't want to make it sound like a negative thing so that people would say,
that's bad. I don't do that.
We just described it in neutral terms.
Sometimes people do this. Do you do this?
And what we found is that people said, oh gee, other people do do that.
You know, that's so great.
You put that into words like that.
I see that all the time.
But me, I know I don't really do that.
So what happened was people recognized the bias
as something that people do,
and they attributed it to other people,
but they thought that they did it quite a bit less.
And the same thing happens in so many different domains.
If you asked me, do you evaluate the news fairly?
Are you a good judge of policy?
I'll tell you, of course I am,
but I can see lots of biases in the people around me.
Emily, you call this the bias blind spot.
What do you mean by the term?
The reason why I came to call it a bias blind spot is that a blind spot refers to a situation
where you can see something sort of all around you except in one place.
And so the blind spot is seeing the bias in yourself because it turns out that people
could readily recognize these biases all around them.
Let's look at some specific domains where the bias blind spot affects us.
When it comes to ethics, we're all quick to see conflicts of interest in other people,
but slow to see it when it comes to ourselves.
It's such a beautiful example.
So you know, doctors and gifts from the pharmaceutical industry.
So people have studied this.
And doctors will say, I'm not influenced by gifts.
And oftentimes the gifts are small, right?
They're like, you have a Pfizer pen.
Sometimes the gifts are rather large.
Like, we'd love to hear you come and give a talk
on your research, you know, in the Caribbean, you know,
we'll fly you over there, you know, in a private jet
and to give your talk.
And I credit to the medical industry that I think, you know,
they have really worked on trying to root this out because they recognized it as a problem.
So there's no longer free lunches for residents every day, you know, sponsored by various drug companies as far as I understand.
But the point being that doctors said that they were not influenced by these gifts, but that other doctors were.
So it's a perfect example of a conflict of interest not being recognized in self but seen in others.
The bias blind spot also affects how we think we are affected by marketing and how we think
others are affected by marketing.
I want to play you a clip for an ad that I recently came by.
GLH means great looking hair.
Just spray GLH on and it instantly covers your ball spot, leaving you with great looking hair.
GLA says not a paint or a cover-up. It's an amazing powder that cleans to the tiniest hairs on your head.
Order GLA now for only a little bit more.
So that was an ad for spray on hair.
Now, I don't think I'm influenced by advertising, whether that's commercial advertising or political advertising,
but I think other people are quite vulnerable to such persuasion.
Yeah, there's a phenomenon called the third person effect whereby people think that persuasive attempts
have more of an impact on other people than themselves. So they say,
oh, commercials, political ads, those things, I'm sort of immune to them.
They don't influence me. Whereas people recognize it's a whole industry.
It's influencing other people, but we see others as more susceptible to these influences
than ourselves. Yeah. How does this work in politics? When we evaluate our political opponents,
how does this bias sort of play out in our evaluations both of people on our side and people on the
other side? Yeah, so that's a great question. And obviously, we've all been thinking about it a lot recently. So I think there's a lot of things that are going on. One is, what do I believe are the
roots of my political opinions and political beliefs? And people will swear that the roots of their
political beliefs are just an irrational analysis of the issue, right? So I take the positions I do
because those are the correct positions. If you analyze the issues, if you analyze the state of the country, if you think about
what's best for the nation, these are the correct positions. But they don't view the,
that as being the root to the positions of those on the other side. So the other side
is influenced by ideology, by self-interest, by prejudice, whatever it is.
Yeah. You know, I was thinking about this, the study that came out some time ago, and this was during the
Obama presidency.
Gas prices were really high, and people were asking, how much is the president responsible
for high gas prices or low gas prices?
And what was interesting is that the same group had asked the question to citizens during
the presidency of George W. Bush when also gas prices were
high. And what's fascinating and perhaps unsurprising is, of course, when gas prices are high
and there's a Republican in the White House, most Republicans think the president has very little
control over gas prices and therefore should not be blamed for it. And Democrats think the
president has a lot of control of a gas price and should be blamed for it. And the tables are exactly
turn when you have a Democrat in the White House.
In both cases, people in very self-interested way see the data that they have and interpret
it in a way that aligns with their political beliefs.
Of course, this is just one example.
There must be hundreds of examples like this.
That's right.
What's amazing is it's motivated reasoning.
Both of those words are important, right?
So it's motivated.
I'm seeing things in a way that's consistent, right,
with my motives, my prior beliefs.
But it's also reasoning, because they're not just saying,
well, I'm just going to believe whatever makes my side look better.
Done, right?
That would be just like pure motivation.
There's reasoning going on.
So people are actually stopping to think,
okay, well, what are the factors that influence gas prices and who might be responsible for? And what's
going on on the global political stage, you know, and and and so if you reason it out,
these things are so complex that you can find reasons for almost anything. Mm hmm. Or at
least for one of the two sides of the issue, anyway.
There is another curious dimension of the bias blind spot
when we come up with positions on various issues
where keenly aware of the nuances and subtleties
of our opinions, but we don't extend the same respect
to the views of our opponents.
We very rarely say the views of the people who disagree
with me are thoughtful and nuanced, right?
I think that's right.
I think that there's more of a tendency to stereotype and caricature
as others and to recognize the nuance and complexity in our own views.
And unfortunately, the political realm, I think, affords that makes that even more likely
because people can't really express their own ambivalences and nuances, right?
Because that's seen as sort of giving into the other side.
So people do tend to portray themselves as sort of more clear
and perhaps even more extreme in that respect as a result.
We've known for a long time that our evaluations of ourselves are very different
than our evaluations of others.
The Bible asks why we notice a speck of dust in our brother's eye, but ignore the beam
sticking out of our own eye.
When we come back, the psychological quirk that produces the radically different judgments
we make of ourselves and others.
You're listening to Hidden Brain, I'm Shankar Vedanta.
This is Hidden Brain, I'm Shankar Vedanta.
All of us find it remarkably easy to identify bias among other people, especially our opponents.
And all of us find it maddeningly difficult to spot biases in ourselves.
Psychologist Emily Pronen has spent years
studying this discrepancy in our perceptions,
and she has found that much of the discrepancy
comes down to the different yardsticks we use
in judging ourselves and others.
Emily, I'm not sure if you have watched
the television show Vipe, but on the show,
there's a character named Jonah Ryan
who decides to run for president.
He's been advised that it's a bad look to be single, so he gets together with a woman
who happens to be the daughter of a man his mom used to be married to.
So she is his step sister.
In an interview, Jonah prefers to think of his fiance as only his former step sister.
So what would you say to someone who might ask how can they marry
their step siblings? I'm not her brother nor have I ever been her brother. Right.
And the only time anyone could ever say that would be for that one year. I mean,
it's exactly what Woody Allen did and nobody thinks he's weird. I mean,
everybody just hates him because Ann swans in as good as a bug slideugs Life. Exactly. So this is obviously a comedy show, Emily.
But I'm wondering if you can just start by explaining when it comes to our judgments
of other people, what is the yardstick that we use to evaluate that they are biased?
The yardstick that we use is, in one word, behavior, their actions.
The process is, it's almost like like if you imagine a fork in the road
and it just goes two different ways there are just two different paths here.
There's the path that we use for self-judgment and there's the path that we use
for judging others. And in my view the path that we use for judging others is
we look at their actions. The path that we use for judging ourselves is we look
inwards. And when I say look inwards, I mean, we look to things like our thoughts,
feelings, intentions, motives.
So if the question is, did I marry my brother?
There's an action, there's a behavior, right?
I did it or I didn't,
and that's how other people will judge it.
But in judging myself, I might look much more
to my motives and my intentions.
Am I someone who would intend to marry their brother?
No, that's weird. I would never intend to do that.
So I guess I didn't do it.
And so, when we are interacting with other people, what we see is them.
We see their actions. We see their expressions.
But when we experience ourselves, we don't really see that.
Instead, what we perceive is what's inside our heads.
That's the information that we're flooded with.
That's the information that we can't escape
is our thoughts and feelings and intentions.
So that's what we give so much weight to.
So one of the things that jumps out at me
from what you're saying, Emily,
is that our introspection is our access to our own thoughts and feelings. These are with us all the time.
So it's almost, we don't actually have to ask ourselves the question, how do I evaluate
myself? We automatically go to looking inward to our thoughts and feelings. When it comes
to our evaluations of other people, in some ways we don't have access to their thoughts
and feelings. Those are hidden from us, and so we use what we have.
And on the surface, this happens without any sort of conscious awareness that it's happening,
right?
So I don't realize I'm using a different yardstick to evaluate your behavior and a different
yardstick to evaluate my.
Right.
I don't think that we really think about that explicitly.
What we use in any judgment, psychologists can tell you, is the information that's salient.
Psychologists like that term salient.
The information that's available to us,
whatever's fresh in our brains is the information we use.
And so it just so happens that for the self,
the information that's fresh in our brains,
all the time is that stuff that we perceive
to be in our brains, right?
Our thoughts and feelings, all that stuff that we perceive to be in our brains, right? Our thoughts and feelings,
all that stuff that's sort of just constantly there and that we're sort of constantly aware of.
It's not just that we use different yardsticks in evaluating ourselves and others. Each
of those yardsticks is flawed and flawed in a different way.
When it comes to evaluating our own behavior through introspection, we imagine that we can see
all our motives and intentions that they are accessible to us. But it turns out that is not the case.
There's a bunch of stuff that goes on in our brand that we're not aware of, right? We're not aware of the sources of our beliefs.
We're not aware of, you know, if I go to the ice cream shop and I choose the chocolate ice cream over the vanilla,
I am aware that that was my choice, but I'm not aware why that was my choice.
That's happening in the brain without my having access to it, but we sometimes forget that.
So we think that we can look inwards and find out everything.
So we forget, for example, that a lot of prejudice and stereotyping happens unconsciously.
And I mean, I can't look inwards to find it.
I'm not necessarily going to have those racist intentions.
That's something that people talk about a lot now. So it's not the case
that just because we have access to all this information in our heads, that that means it's
always going to be probative for making whatever judgment we need to make.
I remember speaking some years ago with the researcher Michael Tesla, he ran an interesting
experiment with Republicans and Democrats and this is back when the country was debating the
Affordable Care Act or Obama Care. And what Michael Tesla did is he presented volunteers with details of the
Affordable Care Act, but he told some of them that the plan had been put forward by President Bill Clinton, a Democratic
President who was white, and he told other volunteers the plan was from Barack Obama, a Democratic president who was black.
So same plan, same details, both put forward by a Democratic president, except that one
was white and one president was black.
And what he found was that both liberals and conservatives were subtly biased by their
feelings about the racial identity of the president.
White racial liberals become more supportive of a policy when it's framed
as Barack Obama than when it's framed as Bill Clinton's, but white racial conservatives
become less supportive of that policy. And Emily, I feel like this is speaking to what
you just said. If you ask liberals and conservatives, how are you evaluating this policy? They will
dive into the details and say, here's why I like the policy or here's why I don't like the policy.
And neither will say my affinity or my aversion to someone from a different race might be
shaping my view on something like the Affordable Care Act.
Yeah, I think that's exactly right.
And that sounds like a great study.
And it's not that the subjects were lying.
They were saying what they believed to be the case. They assumed,
incorrectly, that if the race of the candidate had impacted their judgment that they would know it.
Now an outsider might be able to notice this pattern much more quickly because they wouldn't
be relying on their their intentions. They just be looking at what the person did. Right? So sometimes
when we look at behavior, things can be a little bit easier to see. Yeah. Some years ago, Emily, you came up with a theory of why
our introspection are undrilivable. You called it the introspection illusion. What is the introspection
illusion? Although we have access to our introspection, This is sort of what it means to be a conscious person as you know your thoughts and your feelings
and your motives and your intentions and they're there all the time in your head.
We have some illusions about what that can do for us.
So we think that that gives us sort of supreme self-knowledge, sort of that we can know all
sorts of things about ourselves because we have access to this information.
We also think that our behavior is less important
than knowing what's inside of our heads.
In the case of ourselves, it's our intentions
that are so important to know.
You know, I was speaking some years ago
with Mazarin Banaji, the psychologist at Harvard,
and she said something really interesting to me.
She said, you know, if you have a problem with your heart, you might go to a cardiologist
to get it checked out.
And the cardiologist says, here's what's wrong with your heart.
You are inclined to believe her because you think the cardiologist knows more than you do
about your heart.
You don't tell the cardiologist it's my heart.
Therefore, I must be the expert on my heart
because it belongs to me.
But my serenity was saying the same thing
doesn't happen with our minds.
It's very hard when an expert comes along and says,
let me explain to you how your mind works
because at some level, all of us feel like we are experts
in our own minds.
And that's partly I think connected to what you're calling
the introspection illusion.
It feels like our mental worlds are so rich
and we spend so much time in them
that it feels in some ways we understand how they work.
And in some ways, that could be an illusion.
That's right.
And look, it's embedded in the history of our own field.
In the very early days of psychology,
when people wanted to understand the mind,
they had people come into the laboratory,
sit in a room, and interest back. And they said, this is how we're going to learn understand the mind. They had people come into the laboratory, sit in a room, and
interest back.
And they said, this is how we're going to learn how the mind
works.
And then there was a huge backlash.
The behaviorists came along and they said, this is nonsense.
We cannot learn about how the mind works by asking people to
report to us what's going on in their mind.
So they said, we're getting rid of all of that.
So we're just going to focus on behavior because that is
observable.
And then we had a third wave of cognitive psychology where we realized
that there were objective strategies and parable methods that we could use to study the mind
that did not rely on people telling us what was going on in their own minds.
And I think part of this also rests on the idea that if everything that happens in our
mind was actually accessible to conscious introspection, we might in fact, if we were very honest and very diligent,
be able to look inside our minds and see everything, but in fact, if much of our minds actually
are offering outside of our conscious awareness, what our minds are doing is simply not accessible
to us through introspection.
Right, exactly.
You brought up the halo effect when we were talking earlier.
And the famous experiment on the halo effect
comes from Tim Wilson and Richard Nusbit back in 1977.
And they had people watch a video in which a professor
with a quote unquote foreign accent,
I'm not sure what his accent was talking.
And then they asked subjects in the experiment to evaluate this professor. And half of the subjects had seen the professor
in the video acting cold, not very likable. And the other half of subjects had seen him acting
warm, very likable. And afterwards, they asked the subjects what they thought of him. And
what people said in the unlikable professor condition was that they didn't like his accent
when they were asked about his accent. In the likable condition, they did like his accent.
So what happened was the likability of the professor, which was manipulated by the experimenters,
influenced how much people thought the accent was likable.
But they didn't realize that this had happened at all. They had no access to what had influenced their perception of the accent.
Now, we knew as experimenters, right?
Because we saw those who got the unlikable professor thought it was a bad accent,
and those who got the likable professor thought it was a good accent.
So the experimenters could say, gee, we know how they came to this conclusion.
But the subject didn't know that the subject just looked inwards and said,
that's a bunch of hooy.
Why would I evaluate someone's accent, how likable their accent is,
based on how nice they were that makes no sense?
They had no awareness of having done that.
It occurred unconsciously, and they denied it.
You can see the same thing play out
on a much larger scale when it comes to politics.
In the 2020 US presidential election,
the vast majority of Republicans voted for the Republican candidate
and the vast majority of Democrats voted for the Republican candidate, and the vast majority
of Democrats voted for the Democratic candidate.
Many people, if you ask them, why did you choose the candidate you voted for, they would
give you a nuanced explanation of why candidate A was better than candidate B.
But imagine that you're a neutral observer who just landed in the United States from Mars. You might look at the
same results and say, well, people are just voting for their parties candidate. It doesn't
really make a difference who that person is. Most people tend to dislike such comments
because they suggest that our choices are less deliberate than we think. They prompt
us to search for justifications that prove that we are, in fact,
in charge of our own minds.
That's right.
We think we're being rational
that we're choosing the candidate based on
rational decision making and rational analysis,
but really what we're doing is we're rationalizing.
Really actually, there's other factors
that have determined which candidate we prefer.
And then after the fact, we rationalize it by coming up with what seem like rational
reasons.
And it feels like our evaluations of ourselves and others, you know, is shaped by these
dual forces.
On the one hand, we ascribe greater weight to our own introspection than maybe we should.
But on the other, we discount the introspection of other people.
So in other words, I don't think that my political opponents
have actually thought very carefully about how they've chosen
their course of action.
I sort of can dismiss them as being easily led as being sheep.
Even as we overvalue our own introspection,
we undervalue the internal thought processes of other people.
Yes.
And we've even done experiments where we say, look,
maybe the reason
why people undervalued other's thought processes is that they just don't have access to them,
right? You know, you have such rich access to what's going on in your head. So we will
give people an entire think aloud protocol, meaning that before the subject made their
decision, they thought allowed into a tape recorder. They just dumped all their thoughts and will give that to the subject to hear. And they still disregard it. So we actually did
a study with political beliefs. And this was with Jonah Berger and Sarah Maluki, we made
up various California propositions. One was about increasing the maximum cargo size
at the port of Los Angeles. When then we told people, there are parties positioned on them.
So the Democrats support this or there are Republicans support this.
And we ask them to choose their position to vote essentially.
And we ask them were they influenced by their parties position?
And Subject said, no, I wasn't influenced by that.
I just evaluated the issue.
Before they said what position they would take,
we had them list all their thoughts.
They dumped out all of their thoughts on the issue.
And then we gave that to another person
to also evaluate, and the other person
didn't care about any of that.
So people said, I went with my thoughts,
I evaluated the issue, but the outsider said,
oh gee, I don't need to see all those thoughts,
that's not relevant. You're a Democrat, you want the Democratic, oh gee, I don't need to see all those thoughts. That's not relevant.
You're a Democrat.
You want the Democratic position.
Done.
Simple.
One of the other ideas that's connected to your work on how we overvalue the things happening
in our minds and pay less attention to the things happening in other people's minds
is a phenomenon called naive realism.
Can you talk about what that is and how it connects to your work?
So much of what we're talking about today, I think, is really rooted in the basic functioning
of our brains.
It's just how we are designed, right?
So for example, we have eyes in our head, and those lead us to see other people's behaviors,
but the eyes don't look inwards.
And there's just sort of basic brain architecture that determines so much.
And now you've realized some has to do with this. It has to do with the idea that there are some basic brain architecture that determines so much. And I have realism has to do with this.
It has to do with the idea that there are some basic and inescapable beliefs.
And one is that I believe that I see the world as it is in objective reality.
And as a result, I think that others will see the world the same as I do.
And that when others don't, I have to explain it.
And the way that I tend to explain it is either by saying,
they don't understand, I need to educate them,
or failing that saying, there's something wrong with them.
Either they're stupid, or they're biased.
You can see naive realism at work in everyday interactions.
Take, for example, something that the comedian George Carlin observed.
Have you ever noticed when you're driving that anyone who's driving slower than you is
an idiot? And anyone driving faster than you is a maniac!
Yes, I love that quote. It reminds me of one time I was in the kitchen preparing some food
and my seven year old was in the playroom with my father, his
ground father, and they were looking at different cars in a magazine, and my son kept preferring
the big SUVs, you know, and my father preferred the little boxer-type cars, and at some point
my father said to him, I know what's a matter of taste, but your taste is stupid. That's a great story.
That's a wonderful story.
I feel it speaks to something that I think is really important.
Parents and teachers are constantly trying to teach this lesson.
Don't jump to conclusions, slow down, don't assume you know what's happening in someone
else's head.
And yet it's so hard to remember
to practice these lessons, right?
I mean, as a parent and as a teacher yourself,
do you sometimes go, I'm doing the exact same thing.
I tell my kids not to do.
Yes, because these things are so automatic and so natural.
These are tendencies that we have to override in ourselves.
We can't eliminate them, right?
Because so the tendency to think you see the world as it is, an objective reality.
And therefore, if you like the race car better than the SUV that it truly is better, that
tendency is sort of an escapeable.
And when children do it, they don't realize even that there's a distinction when they're
young between their perception as reality. As we get older, we come to recognize, right, oh, wait a second, that's a distinction when they're young between their perception is reality.
As we get older, we come to recognize, wait a second, that's a matter of taste.
At some level we come to realize, oh no, there's different perspectives and it's a
martyr.
But that initial belief that we have from childhood that my perception is reality doesn't
really go away.
And so it does actually feel that the car we prefers the better one.
There are lots of implications that stem from what Emily calls this basic
architecture of the brain. Here is one that should be familiar to all of us.
If I'm late for a meeting, my mind is chock full with all the reasons I'm late.
Traffic was terrible, I had a childcare crisis and so on.
But if someone else is late for a meeting, I don't have access to all that stuff happening
inside their heads.
It's easy for me to think of them as just being irresponsible or careless.
Psychologists call this the fundamental attribution error.
Another implication of this work has to do with the phenomenon of magical thinking.
Magical thinking involves the idea that our thoughts could somehow influence the world around us.
So, for example, if I think ill thoughts about you, can that give you a headache?
Or if I think positive thoughts about my favorite player
on the team, will that help them score a goal?
And we found that, in fact, this was the case.
So for example, we had people think evil thoughts about
someone else in the experiment.
We said we're interested in whether you could place a hex
on the person and then they stuck pins in a voodoo doll.
And then the other person reported a headache
because they worked for us and we told them to.
And what happened was if you were told to think ill about the other person before putting
those pins in, you were told, just take a minute and think of something terrible. Just think of the
worst thing you can happening to this person. And then you stuck the pins in the doll. Then you
felt like you caused the headache and you felt bad. And we found the same thing with basketball.
Before a big university basketball game,
we had people think about the different players
and think about how each one of them could contribute
to the game and how would they help their team score well.
And then we asked them after the game how much impact
they felt that their thoughts had on the score of the game.
So they thought that they'd impacted the game
when they had thought about the players
doing well. And the way that it's related is it again involves putting too much weight
on what's going on inside our heads. Because we're basically saying that what's going
on inside my head could give someone else a headache or what's going on inside my
head when I sit in the stands of a basketball game could influence the players'
score.
How many baskets they shot.
I mean, if it's a critical moment in the game, I would feel terrible getting up to leave
the room and get some popcorn.
I feel like, I can't let my team down.
How could I do that to them?
The introspection illusion, the biased blind spot and naive realism have profound consequences
in our daily lives.
They do more than shape our thinking in basketball games.
They shape life and death decisions and choices to go to war.
That's when we come back.
You're listening to Hidden Brain, I'm Shankar Vedanthan.
Psychologist Emily Pronen has found we judge ourselves very differently from the way we judge
others.
This is because we use different yardstakes while doing those two things. We evaluate others based on their behavior, but we evaluate our own actions using introspection,
and it turns out introspection is not a useful guide to understanding our own minds.
Emily, I want to talk about some of the implications of this work and the ways in which it plays
out in the real world, and I want to start with an example of something
that can seem trivial, but that produces widespread conflict
across the United States.
I'm going to let Wuppie Goldberg explain.
A survey found that one of the most common arguments
this time of year in households across America
is what temperature to put thermostat at?
Right.
Everybody relates. Oh my gosh. I mean, you are like, uh-huh? Yeah. Everybody, everybody, we race.
We are like, uh-huh.
Yeah.
There's a lot of stuff we should be talking about,
because it's on the list, but this interest made.
Because I feel like everybody deals with this,
if you're interested.
So Emily, I want to draw attention to the fact
that this is a topic which you bring up.
Almost everyone has an opinion about it and the
opinion is often heated. Tell me how it connects to the conversation we're having about how we think about our minds, other people's minds, and the
judgments we arrive at.
I just love the idea of the thermostat wars is it's just it's so real.
And I think it goes back to that quote. I know it's a matter of taste, but your taste is stupid.
Because essentially, if I think that the temperature should be set to 73 and you think it should be set to 68,
I do realize that this is a matter of taste, right? That there's no right answer here. But at another level,
I actually think that the temperature that I wanted to be at is the correct one.
And it's like that George Carlin quote, right?
If you want it to be hotter than me, you're a little soft and ridiculous,
you know, and you're wasting a lot of energy.
If you want it to be colder than me, you've got to be kidding.
Do you really need to be that aesthetic and suffer like that?
We can turn up the heat a little bit more, right?
So we think it's a matter of taste, but we also think we're right.
And I've done some research with Nate Sheik and Shane Blackman
where we show this with paintings.
People say, oh, paintings, that's a matter of taste.
Until someone disagrees with them about which are the nice paintings
and which are the bad paintings.
And then all of a sudden, they say that that person is wrong.
Tell me a little bit more about the art study.
I'm fascinated by the examples that you used and what you found.
So Shane collected some images of paintings from art history books. So these were paintings by famous artists.
And they were in major museums. And they were arranged from abstract to portraits, and we would show them to a subject and ask them to
rate which ones they thought were truly great and which ones they thought were overrated.
And then we showed them cover story for this of another subject who had supposedly done
the same task, but that subject totally disagreed with them.
So if I thought it was great, they thought it was overrated and vice versa.
And then they had to evaluate this other person. And although they said that opinions on art were a matter of
taste, when they saw this person who disagreed with them, they actually thought
that the person was wrong and had been influenced by improper influences,
because otherwise they surely would have agreed with oneself. And we've done
studies with chefs, Kobe Sassarski, my student, did a study with chefs
and chefs showed this exact phenomenon, right?
There's an objectively correct amount that the meat should
be cooked, that the pasta should be cooked,
you know how much it should be salted,
and those who do it the other way are wrong.
I want to point to something that you said earlier that I think might connect with this,
which is in some ways, when we think about our own subjective conclusions, when we think
about a painting or how long past I should be cooked, we're actually not thinking of this
as being subjective. It genuinely feels as if we have a master, a whole bunch of objective
data, and a right to this conclusion that in some ways feels objective.
So, you know, we might say, yes, my taste in art and music is subjective, but it actually feels like it's not that it's actually objective.
And part of it is that when things come to us through the senses, it comes so quickly that we do not feel the operation of the mind being involved. I know if I preferred to get the chocolate ice cream to the mint chip ice cream, but if
you ask me why, I simply don't have access to that.
And so it doesn't feel like there's been all these intervening processes that could have
biased it.
So when the piece of food hits my mouth and I think that it has too much salt, I don't
have access to any brain processes
that are influencing that judgment. It's just, yeah, that's too salty. And it's an immediate
feeling and because it is so immediate, it's hard to imagine that it could have been biased
by anything.
I'm wondering if this is connected in some way, Emily, to other work that you have done
that looks at how we perform during interviews, but also how we judge other people during interviews.
So we all think that we can sit before someone
for half an hour, talk to them,
and get a pretty good sense of whether this person
is a good fit for a job.
But if someone were to come along and say,
oh, we can talk to you for half an hour,
and figure out if you are a good fit for the job,
we say, that's clearly inadequate
because I'm so much more complex
than anything that can be ascertained in 30 minutes.
Right.
So there's a term, the interview illusion, which I did not coin, and it was about this
idea that it's an illusion that you can tell so much from an interview.
If I want to know whether you're going to be a good brick layer, it's probably a lot
more valuable for me to watch you lay bricks and for me to ask your five prior bosses, how well you lay bricks, then for me to sit down
and interview you about how good a brick layer you are.
And yet people love interviews.
And even psychologists, we do job interviews.
Why do we do that?
We could just ask the people who've worked with the person, you know, we could ask their
advisors to write letters and we could read their work but we do interviews as well
and we put a lot of weight on them and the work that I did we actually had people come into the
laboratory and pairs these were students who never met each other and they talked to each other
for a half an hour and we found that at the end of the half an hour they felt like they'd really
come to know the other person but that the other person had not really come to know them so you
could only get a small understanding and a small glimpse of who I am from that conversation.
But I've got the whole you.
And part of that has to do with the fact that I know all the stuff about me that you didn't
find out from that conversation.
I'm aware of all the stuff I didn't say.
All the stuff I said that maybe was misleading about who I really am.
I've got all that.
But I don't have all that about you.
You can see how these biases might play out in the context, not just of interpersonal conflict, but geopolitical conflict.
If you think you see the world accurately and I don't,
if you try to set me straight and find you can't change my views,
what are you to conclude?
The simplest explanation is that I can't be trusted.
There's no point trying to understand me, or reason with me, or negotiate with me, because
I must be either stupid or evil.
It's not just people's actions that influence how we want to respond to them.
It's also our beliefs about what those actions stem from, and if we believe that individuals are biased, that their mental processes are biased, then we don't believe
that it makes sense to try to reason with them. Is there any evidence that teaching people about
the ways in which our minds work, that it actually changes the way they can actually perceive the conflict and perhaps respond differently.
When people learn about these different biases, they're initially very optimistic that what
we need to do is educate people about the biases.
So if I just tell people, like my students, here's the different biases that people engage
in, that that should solve the problem.
They'll say, gee, I didn't know about all those biases.
Yeah. And the idea is now that I know about them, I won't do them.
But as you know, that's not how it works.
Because what happens is they say, gee,
I didn't have words for all those biases.
But now that you've told me the words,
they give me a great vocabulary for describing
what all these people around me keep doing.
So that doesn't work.
That doesn't work.
But what not you cooked there, and I tried,
was instead to educate people about the importance
of unconscious processes.
And we taught people about how a lot of our judgments
are rooted in things we don't have access to,
so that a lot of things are automatic,
and a lot of things are biased.
And so we tried to educate them about essentially
the introspection illusion,
the illusion that we have access to all these things, and the fact that instead much of
it is occurring automatically and is biased, and then we asked people to complete our usual
bias blind spot measure where they read about various biases. And then people no longer
showed a bias blind spot. So once they understood about the operation of the unconscious
and how these things happen automatically,
they no longer claimed to be less biased than others.
So then they said, gee, maybe I am biased,
maybe looking inwards and not seeing bias
is not the best way to conclude whether I'm biased or not.
Can you give me a concrete example of a time
when you used your own research to change how you thought about something important or to change your own behavior.
I don't know if I can give you a single important example, but I think that as a parent, I find it happening with me all the time.
That, you know, I'm talking to my kids about someone in our lives who's done something that sort of has irritated us in some way.
Somebody canceled on a plan that we had or said something that was insensitive.
I find myself doing that thing where I'm about to jump to the fundamental attribution
or I'm about to say, gee, that was mean or inconsiderate or lazy or whatever, and then
I've got my kids there with me. Oh, this is not what I want to teach my children.
And so I say, wait a second, you know, I know it might seem like the person was being
inconsiderate, but maybe they're having a really hard day.
I tried to sort of teach them and to remind myself to think about people's circumstances
instead of jumping right away to that dispossitional attribution.
You know Emily as I'm thinking about your work, I'm realizing that many miscommunications
might happen because our thoughts seem so clear to us, but we do a terrible job communicating
those thoughts to others.
Things in our minds seem so clear and loom so large to us that we somehow assume they
must be clear to others as well.
Yeah, it's interesting. I think of the example of breakups, romantic breakups, and people
sort of they want to be kind and they want to be considerate and they want to do it nicely
not always. And the other person is left totally confused. And he's, oh, you know, yeah,
I think, you know, we're just taking a break for a few days. We just hit a rough patch.
The other person thinks they have successfully broken up and ended their relationship.
We forget that our intentions, what we're intending to do is to break up, to close the door,
but to do it in a very kind and considerate way.
The other person thinks is that you've sent a bunch of mixed messages and you're leaving
the door open. So this is just one of so many examples
where we don't recognize our lack of transparency
and what we haven't communicated
because it's so obvious to us.
I don't know if you're a fan of the show Parks and Recreation,
but there was an incident on Parks and Rec
that basically is almost exactly the same line,
so one party is breaking up with another,
but they break up so politely
and so kindly
of the other party thinks, great, we've had a wonderful chat, the relationship is now
on to a higher level than it was before, and one party thinks they're broken up with the other party
things. Wow, we're really on a good place now. So you're leaving soon. Back to Indianapolis briefly,
and then on to a town called Snirling Indiana for several months.
Never heard of it.
It's quite small.
The cows at number the people 41.
And then after Chris moves and tracks him down, storms into his house and accuses him of
cheating on her.
God.
I'm so sorry, honey.
I'm so embarrassed.
I was scared that you were cheating on me.
Oh, I'm cheating on you, but I'm also not dating you.
We broke up last week.
Yeah, I'm laughing, but it's actually sad.
I mean, it actually causes a lot of suffering and reality.
Yeah, we've just been through a really bruising political year, Emily, and the country as
a whole has been very divided.
And a lot of people are really asking asking is it possible for us to come together
as a nation after you know very bitter political fight. I'm wondering if you
were to give advice to the nation based on the work that you have done what would
that advice look and sound like? I mean I think one thing I would say is that if
you were judging yourself by all your
positive intentions, right, and you're good feelings, you know, if your intentions are
that you want the country to be in a better place, that you want people to thrive, don't
assume that others intentions are different from your own. And if you put a lot of weight on your intentions, those other intentions would deserve just that
same amount of weight.
So we owe some charity in judging others' behavior by giving some weight to their intentions,
and we should not assume that their intentions are so different from our own.
And if we can start with that and start with the charity of trying to find the positive
intentions and others, then maybe there is some hope, but it's so hard to do, especially
when things are so divided and feel so divided.
Psychologist Emily Pronen teaches at Princeton University. Emily, thanks for joining me today on Hidden Brain.
Thank you so much for having me.
Hidden Brain is produced by Hidden Brain Media.
Our production team includes Bridget McCarthy, Ani Murphy-Paul, Kristen Wong, Laura Correll,
Ryan Katz, Autumn Barnes and Andrew Chadwick.
Tara Boyle is our executive producer.
I'm Hidden Brain's executive editor.
Our unsung hero today is HR specialist Max Lenowitz.
When we launched Hidden Brain Media, Max was an onboarding manager at JustWorks, the company
that handles our payroll.
Max worked closely with us as we were launching our business, and he patiently answered our
many questions about how to pay employees and track things like vacation and sick time.
His good cheer made a busy time feel less stressful.
Thank you Max for helping us to find our feet.
For more of our work please be sure to subscribe to our newsletter.
Every week we feature interesting research on human behavior and a brain teaser.
Plus, we give you a look at what's coming up on both Hidden Brain and our new show My Unsung Hero.
You can subscribe at news.hiddenbrain.org. That's N-E-W-S at HiddenBrain.org.
Next week in our mind reading series, we look at how many of us underestimate
the goodness in human nature.
Guess what people aren't entirely selfish economists have to learn that lesson too.
People aren't entirely selfish we actually care a lot about others.
I'm Shankar Vedantam. See you soon. you